-
Notifications
You must be signed in to change notification settings - Fork 1
Added missing sentiment classifier, fixes #1 #2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
- Added API folder - Train and Test notebook for each part of the process, to break it up
- Cleaned up imports
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you for moving the code around to the api folder. Only a few minor changes. When you looked at the original file, were there pip commands somewhere?
I am not seeing anything that install the dependencies. Could you create an env.yml file here without using the export? (see my remark on env.yml file on the process mining project)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can you combine this to test.py and rename to model.py?
| "id": "RAXtSnSK4LPr" | ||
| }, | ||
| "source": [ | ||
| "# Tokenlized\n", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Typo?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
rename file to tokenize_statistics.ipynb
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Note to self: This notebook does not tokenize anything currently, it just adds a column with 1s associated to the tokenizer. Not sure why, will figure out in the future.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We confirmed that the only files that will trully be needed on this notebook are:
- so-dataset.csv
- gh-dataset.csv
- crossplatform_sf_dataset.csv
and also the one column ones:
- so-dataset_tokenized.csv
- gh-dataset_tokenized.csv
- crossplatform_sf_dataset_tokenized.csv
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Note to self: This filter.py functions are not used in the entire original notebook, and therefore it is not used in these 3 refactored notebooks.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Note to self: This tokenizer.py functions are not used in the entire original notebook, and therefore it is not used in these 3 refactored notebooks.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Note to self: This notebook does not tokenize anything currently, it just adds a column with 1s associated to the tokenizer. Not sure why, will figure out in the future.
|
So just to confirm I just need your pips if possible as a env.yml and to merge the api files of train.py and test.py to model.py |
- train.py and test.py now exist in model.py - env added for required packages - Minor typo changes Signed-off-by: Connor Narowetz <cnarowetz@gmail.com>
- __init__.py added - docs for model.py added Signed-off-by: Connor Narowetz <cnarowetz@gmail.com>
Pdocs Attached |
|
@connorn-dev thank you for remembering this! |
No description provided.