-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Cannot find embeddings #3
Comments
Hi! It seems there's an issue with the path of the embedding file. Could you check two things:
|
Thank you very much for the quick response! Unfortunately, I could not resolve the issue. After restarting my device, the path and file was found, however the programm tells me, that the embeddings file contains errors and I am unsure how to deal with that.
|
Can you re-download the embeddings file making sure it is downloaded properly? (it seems the file is broken there). Note that the file size should be around 1.3G |
Yes, it is downloaded correctly and 1.3GB is also correct.
Federico Nanni ***@***.***> schrieb am Di. 16. Jan. 2024 um
12:58:
… Can you re-download the embeddings file making sure it is downloaded
properly? (it seems the file is broken there). Note that the file size
should be around 1.3G
—
Reply to this email directly, view it on GitHub
<#3 (comment)>, or
unsubscribe
<https://github.com/notifications/unsubscribe-auth/AUAYD6OD7BUNG6EF2DFFSXLYOZTQDAVCNFSM6AAAAABBYHNVF2VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTQOJTGYYDANJXGY>
.
You are receiving this because you authored the thread.Message ID:
***@***.***>
|
From here it is a bit hard to debug. I have just reinstalled it all and it seems to be working for me using that input embedding file and the textual data from the online appendix I'm tagging @irehbein because she might be working on Windows on this (I've just tested on Mac and Linux and in both cases it loaded embeddings just fine). Sorry, but it has been a long time since we last worked on this! |
Ah - check the order of the commands! You should have:
Your examples has embeddings first and input folder second:
|
Thank you so much! This is working now. However, I am still having some
issues with the application. I want to use semscale for a csv datafile,
that I have containing tweets of German parliament politicians. Since it
contains tweets of many years, I do have about a million txt files now.
Therefore, I tried running the code a few times now, but it seems that due
to memory limitations it is never able to finish. Do I see it correctly,
that I need a txt file of every tweet in the beginning, starting with "de
/n (text)"? And do you have any advice on how I could use the package more
efficiently?
Am Di., 16. Jan. 2024 um 13:25 Uhr schrieb Federico Nanni <
***@***.***>:
… Fixed it - let me know if this works now:
Screenshot.2024-01-16.at.12.25.24.png (view on web)
<https://github.com/umanlp/SemScale/assets/8415204/f0c2c767-8c9b-4eb1-8aaa-d5bd1ff86feb>
—
Reply to this email directly, view it on GitHub
<#3 (comment)>, or
unsubscribe
<https://github.com/notifications/unsubscribe-auth/AUAYD6MMMDNLZNATB3FTGVTYOZWVHAVCNFSM6AAAAABBYHNVF2VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTQOJTGY2DENBSGY>
.
You are receiving this because you authored the thread.Message ID:
***@***.***>
|
I see, maybe you could group tweets together by author to reduce the number of files. So one file for each user - this way you'll be scaling users, not single tweets |
Hi, thank you so much for providing this code! Unfortunatetly I am having issues running SemScale. In Anaconda Promptshell I ran:
python scaler.py C:\Users\SemScale\embeddings\wiki.big-five.mapped.vec C:\Users\SemScale\datadir_test C:\Users\SemScale\output.txt
However, this always yields the error:
WARNING:tensorflow:From C:\Users\Documents\Python\envs\semscale\lib\site-packages\keras\src\losses.py:2976: The name tf.losses.sparse_softmax_cross_entropy is deprecated. Please use tf.compat.v1.losses.sparse_softmax_cross_entropy instead. Error: File containing pre-trained word embeddings not found.
Is the embedding not working anymore? Thank you very much in advance for your help!!
The text was updated successfully, but these errors were encountered: