Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
45.5
TFLOPS
160
13
62
Jean Louis
JLouisBiz
Follow
jorgemunozl's profile picture
cp-execsearch's profile picture
tensiondriven's profile picture
76 followers
·
114 following
https://www.StartYourOwnGoldMine.com
YourOwnGoldMine
gnusupport
AI & ML interests
- LLM for sales, marketing, promotion - LLM for Website Revision System - increasing quality of communication with customers - helping clients access information faster - saving people from financial troubles
Recent Activity
reacted
to
mitkox
's
post
with 👍
1 day ago
Got to 1199.8 tokens/sec with Devstral Small -2 on my desktop GPU workstation. vLLM nightly. Works out of the box with Mistral Vibe. Next is time to test the big one.
replied
to
mitkox
's
post
1 day ago
Got to 1199.8 tokens/sec with Devstral Small -2 on my desktop GPU workstation. vLLM nightly. Works out of the box with Mistral Vibe. Next is time to test the big one.
replied
to
melvindave
's
post
1 day ago
Currently having a blast learning the transformers library. I noticed that model cards usually have Transformers code as usage examples. So I tried to figure out how to load a model just using the transformers library without using ollama, lmstudio, or llamacpp. Learned how to install dependencies required to make it work like pytorch and CUDA. I also used Conda for python environment dependencies. Once I got the model loaded and sample inference working, I made an API to serve it. I know it's very basic stuff for machine learning experts here in HF but I'm completely new to this so I'm happy to get it working! Model used: https://huggingface.co/Qwen/Qwen3-VL-8B-Instruct GPU: NVIDIA GeForce RTX 3090 Here's the result of my experimentation
View all activity
Organizations
JLouisBiz
's models
2
Sort: Recently updated
JLouisBiz/Qwen3-0.6B-GGUF-for-24GB-VRAM
Text Generation
•
0.8B
•
Updated
Apr 29
•
66
•
2
JLouisBiz/Qwen3-1.7B-for-24GB-VRAM
Text Generation
•
2B
•
Updated
Apr 29
•
9