Martial Terran
MartialTerran
AI & ML interests
I, Martial Terran am leading a Group to build solar-powered TimeCapsuleTeacher(TM} GPT-powered laptop computers, to provide Language, Math and Science Education to Non-English-Speaking people of the future in a Post-Apophis World.
Recent Activity
updated
a model
about 12 hours ago
MartialTerran/Method_for_Dynamically_Reducing_Logit_Computation_in_LLMs
New activity
about 12 hours ago
Qwen/Qwen2.5-Coder-1.5B
updated
a dataset
4 days ago
MartialTerran/Korean_Faces
Organizations
MartialTerran's activity
Optimizing Qwen Coder Models (1.5B & 3B) for Python and Edge Deployment
#6 opened about 12 hours ago
by
MartialTerran
Duplicates in Train set
1
#12 opened about 1 year ago
by
Qilex
Storing Spelling information in LLMs
2
#2 opened 14 days ago
by
MartialTerran
Cleaned data
3
#15 opened 11 months ago
by
ad8e
Request Fork with Modifications for Python GenAI App Development on Microsoft OS
#5 opened 10 days ago
by
MartialTerran
finetuning
3
#2 opened 22 days ago
by
HassanStar
Using Adapter/PEFT for finetuning a Subnet extracted from the SmolLM2 for Arduino Tool Calling
1
#7 opened 18 days ago
by
MartialTerran
Extracting an optimized Arduino Tool-Calling Subnet from the SmolLM2 model.
#6 opened 18 days ago
by
MartialTerran
Extracting subnets from the published SmolLM2 model for compute-efficient task performance on edge devices
#5 opened 18 days ago
by
MartialTerran
Porting SmolLM2 to Arduino
1
#4 opened 18 days ago
by
MartialTerran
Pure C++ version of the SmolLM2 model code for EDGE implementations
#3 opened 18 days ago
by
MartialTerran
Pure Python version for local Inference operation on PC
#2 opened 18 days ago
by
MartialTerran
Can Huggingface facilitate experimentation with Tiny LLMs
5
#2 opened 6 months ago
by
MartialTerran
Link to the Python Script or C-compiled Code to Inference-mode run the Model checkpoint?
#1 opened 6 months ago
by
MartialTerran
Too much Junk vocab words in the vocab.json.
8
#28 opened 8 months ago
by
MartialTerran
The only used vocabulary words/tokens in this model are the letters of the alphabet?
26
#1 opened 8 months ago
by
MartialTerran
GPT-2 model having16 4-float attention heads
#2 opened 7 months ago
by
MartialTerran
Performance oddities of the 3M model.
#3 opened 7 months ago
by
MartialTerran
Model Has Some Coherence. But only uses single-letter tokens?
1
#2 opened 8 months ago
by
MartialTerran
Mismatched Vocab.json verus Words actually within the
6
#19 opened 8 months ago
by
MartialTerran