Building a RL agent "Digital Analyst" for the Indian Stock Market

Hi Groq Community! :waving_hand:

I am building an open-source, news-aware Reinforcement Learning (RL) agent specialized for the Indian Nifty 500 market.

My architecture uses a “Teacher-Student Distillation” workflow:

  1. Teacher: Llama 3.3 70B (Versatile) on Groq provides a “Ground Truth” expert analysis for 479 stocks based on technical chart patterns and live news sentiment.

  2. Student: A lightweight Qwen-0.8B model is being fine-tuned to mirror these expert rationales for ultra-low latency inference.

The Challenge: Your LPU speeds are incredible (300+ TPS), which is exactly what this agentic flow needs. However, I’ve hit the 100K Token-Per-Day (TPD) limit on the Free Developer Tier. To generate a high-quality historical dataset for all 479 stocks, I need to process roughly 5M+ tokens.

Does anyone have advice for:

  • Pacing massive batch-inference runs for “burst” dataset generation?

  • The current status of the Developer Tier upgrade process (mine is showing as closed)?

  • Whether there is a program for hackathon participants looking for temporary limit increases for research purposes?

You can check out the project progress here: [https://github.com/harshgmx47/openenv-nifty500-stock-agent-rl\]

Really enjoying the Llama 3.3 70B performance so far—the latency is unmatched!

#FinanceAI #Llama3 #GroqSpeed #Nifty500 #ReinforcementLearning

1 Like