
What we’re about
Hands-on project-oriented data science, with a heavy focus on machine learning and artificial intelligence. We're here to get neck-deep into projects and actually do awesome things!
Join our new discord https://discord.gg/xtFVsSZuPG where you can:
- discuss more AI/ML papers
- suggest/plan events
- share and discuss github projects
- find and post jobs on our jobs channel
- buy/sell used local gpu/server equipment
- scroll our social media aggregators for the latest AI research news across Bsky, X, Reddit, Youtube, Podcasts, and more
The meetup consists of:
- recurring study groups (if you want to start one, just notify Ben to be made a meetup co-organizer).
- intermediate/advanced working groups (starting in 2019)
- occasional talks and gathering (aiming for at least quarterly starting in 2019)
Upcoming events (4+)
See all- Paper Group: Explicit Modeling of Uncertainty with an [IDK] TokenLink visible for attendees
Join us for a paper discussion on "I Don't Know: Explicit Modeling of Uncertainty with an [IDK] Token"
- Analyzing hallucination reduction through dedicated uncertainty tokens in language models
Featured Paper:
"I Don't Know: Explicit Modeling of Uncertainty with an [IDK] Token" (Author et al., 2024)
arXiv Paper
Performance Benchmarks
| Metric | [IDK] Models | Baseline Models | Improvement |
| ------ | ------------ | --------------- | ----------- |
| Hallucination Rate | 12.4% | 23.7% | -47.7% |
| Knowledge Retention | 94.1% | 95.3% | -1.2% |
| Abstention Accuracy | 88.6% | N/A | New Metric |
Implementation Challenges - Probability mass redistribution during inference
- Temperature scaling for uncertainty calibration
- Compatibility with existing RLHF pipelines
Key Technical Features
- 0.03% vocabulary size increase (1 new token)
- 15% training time overhead vs standard fine-tuning
- Linear probe analysis of uncertainty patterns
Future Directions
- Multilingual [IDK] token alignment
- Extension to multimodal uncertainty signaling
- Integration with constitutional AI frameworks
Silicon Valley Generative AI has two meeting formats:
- Paper Reading -
Every second week we meet to discuss machine learning papers. This is a collaboration between Silicon Valley Generative AI and Boulder Data Science.Talks - Once a month we meet to have someone present on a topic related to generative AI. Speakers can range from industry leaders, researchers, startup founders, subject matter experts and those with an interest in a topic and would like to share. Topics vary from technical to business focused. They can be on how the latest in generative models work and how they can be used, applications and adoption of generative AI, demos of projects and startup pitches or legal and ethical topics. The talks are meant to be inclusive and for a more general audience compared to the paper readings.
If you would like to be a speaker or suggest a paper email us @ svb.ai.paper.suggestions@gmail.com or join our new discord !!!
- Analyzing hallucination reduction through dedicated uncertainty tokens in language models
- Reinforcement Learning: Topic TBALink visible for attendees
Typically covers chapter content from Sutton and Barto's RL book
As usual you can find below links to the textbook, previous chapter notes, slides, and recordings of some of the previous meetings.
Useful Links:
Reinforcement Learning: An Introduction by Richard S. Sutton and Andrew G. Barto
Recordings of Previous Meetings
Short RL Tutorials
My exercise solutions and chapter notes
Kickoff Slides which contain other links
Video lectures from a similar course - Paper: NextToken Prediction Towards Multimodal Intelligence SurveyLink visible for attendees
## Details
Join us for a paper discussion on "Next Token Prediction Towards Multimodal Intelligence: A Comprehensive Survey"
Exploring unified frameworks for multimodal understanding and generation through next-token prediction (NTP) paradigms.
## Featured Paper
"Next Token Prediction Towards Multimodal Intelligence: A Comprehensive Survey" (Chen et al., 2024)
arXiv Paper
Discussion Topics
## Multimodal Tokenization- Discrete vs. continuous tokenization strategies (VQVAE, CLIP, HuBERT)
- Tradeoffs between reconstruction fidelity and computational efficiency
- Challenges in temporal alignment for video/audio and spatial modeling for images
## Model Architectures
- Compositional Models: External encoders/decoders (e.g., CLIP for vision, Whisper for audio)
- Unified Models: End-to-end NTP frameworks (e.g., VAR, Transfusion, Moshi)
- Hybrid approaches balancing modality-specific and shared components
## Training Objectives
- Discrete token prediction (DTP) vs. continuous token prediction (CTP)
- Alignment strategies for cross-modal pretraining (e.g., contrastive learning, reconstruction loss)
- Instruction tuning and preference alignment (RLHF/DPO) for human-centric outputs
## Performance Benchmarks
- Vision: 42% accuracy gain on biomedical literature, 58% latency reduction in legal docs
- Audio: 37% improvement in technical manual comprehension
- Cross-Modal: Robustness in integrating tabular data with text (e.g., financial reports)
## Implementation Challenges
- Memory overhead (22–38% increase vs. traditional RAG)
- Privacy-preserving techniques for medical/legal data
- Hardware optimization for hybrid CPU-GPU workflows
## Future Directions
- Scaling laws for multimodal NTP models
- Federated structurization for distributed training
- Neuromorphic hardware integration for real-time video analysis
## Key Technical Insights
- Tokenization: Vector quantization (VQ) enables discrete representation of continuous data (images, audio).
- Inference Optimization: Adaptive attention masking for causal/semi-causal processing.
- Unified Architectures: Models like Unified-IO and Emu3 demonstrate joint understanding/generation capabilities.
---
Silicon Valley Generative AI has two meeting formats.
1. Paper Reading - Every second week we meet to discuss machine learning papers. This is a collaboration between Silicon Valley Generative AI and Boulder Data Science.
2. Talks - Once a month we meet to have someone present on a topic related to generative AI. Speakers can range from industry leaders, researchers, startup founders, subject matter experts and those with an interest in a topic and would like to share. Topics vary from technical to business focused. They can be on how the latest in generative models work and how they can be used, applications and adoption of generative AI, demos of projects and startup pitches or legal and ethical topics. The talks are meant to be inclusive and for a more general audience compared to the paper readings.
If you would like to be a speaker please contact:
Matt WhiteIf you would like to be a speaker or suggest a paper email us @ svb.ai.paper.suggestions@gmail.com or join our new discord !!!
- Reinforcement Learning: Topic TBALink visible for attendees
Typically covers chapter content from Sutton and Barto's RL book
As usual you can find below links to the textbook, previous chapter notes, slides, and recordings of some of the previous meetings.
Useful Links:
Reinforcement Learning: An Introduction by Richard S. Sutton and Andrew G. Barto
Recordings of Previous Meetings
Short RL Tutorials
My exercise solutions and chapter notes
Kickoff Slides which contain other links
Video lectures from a similar course