Cleaned up code, added multiseed training wrapper, PyTorch profiler training option, updated gradio demo, made changes to research paper to match new changes and new training results from adding new training techniques, architecture.md now explains all designs and decisions
90a2698
OliverPerrincommited on
Added many new improvements based on feedback from others
0d858b5
OliverPerrincommited on
Full training results & evaluation with BERTScore
1e95f87
OliverPerrincommited on
Medium training run
b93250a
OliverPerrincommited on
Fixed compiling issue, added legnth penalty, and atttempting freezing encoder layers 0-5 to lower parameters and preserve T5's langauge understanding.
baf3026
OliverPerrincommited on
Improve emotion task architecture and fix visualizations
6a7a381
OliverPerrincommited on
Clean up codebase and fix training bugs
1601799
OliverPerrincommited on
Redesign Gradio demo with book/news browsing, update .gitignore
fc64ea0
OliverPerrincommited on
Update Gradio demo, inference factory, and evaluation results
e3d3925
OliverPerrincommited on
Update LexiMind: improved training, model architecture, and evaluation
076bc18
OliverPerrincommited on
Improve summarization output quality
69b8f98
OliverPerrincommited on
Fix Pylance type errors, add inductor compilation support
cf79a6c
OliverPerrincommited on
Full training run, code cleanup, mypy/ruff fixes
a47e5cf
OliverPerrincommited on
Training run: dev config 1 epoch results
7bfcb3e
OliverPerrincommited on
feat: Add FLAN-T5 compatibility with relative position bias
Implemented the following parts for the Transformer model: poisional_encoding, feedfoward, encoder, and skeleton of decoder. As well as Test cases for each class respectively and visualizations