- The Renaissance Times
- Posts
- Gemini 1.5's Context Length is a Game Changer
Gemini 1.5's Context Length is a Game Changer
Welcome Renaissance Creator,
Every day it comes time for another AI company to change the game, and today those are our friends at Google.
While we have mentioned Gemini in previous editions of this newsletter, today is different because we have actual experiments we can display from people using the product.
The biggest deal here seems to clearly be the larger context length Google allows as input (you can upload larger file sizes for the model to use in its responses).
This simple improvement has let to a world of experimentation and could be a game changer for certain industries.
Let’s dive in.
Hot Off The Press
Bioptimus: Aiming to build first universal biology AI model
Duet in the workspace is Google’s new AI companion
Generative AI Chip Designer Recogni Locks Up $102M
Microsoft is developing server technology to make AI chips run faster
Adobe Acrobat adds generative AI to ‘easily chat with documents’
Acer’s latest Swift laptops have AMD 8040 chips with Ryzen AI support
Gemini is performing really well on large input tests thus far
The One Big Thing
As simple as it sounds, it turns out that increasing the context window of an AI model (increasing the amount of information it can take in and process) does wonders for its overall capabilities and usefulness.
Gemini impressively boasts a context window length of 1 million tokens in production, a massive increase from the 32,000 limit in Gemini 1.0. According to Google, “This means 1.5 Pro can process vast amounts of information in one go — including 1 hour of video, 11 hours of audio, codebases with over 30,000 lines of code or over 700,000 words.”
To this point, Gemini is already doing things that other AI’s have failed to do for many months and years now:
I showed Gemini 1.5 Pro the ENTIRE Self-Operating Computer codebase, and an example Gemini 1.5 API call.
From there, it was able to perfectly explain how the codebase works...
and then it implemented itself as a new supported model for the repo!
Not perfect, but very close. twitter.com/i/web/status/1…
— Matt Shumer (@mattshumer_)
4:57 AM • Feb 20, 2024
The implications of such a model with a context length of 10 million tokens (their research limit) are profound. This could very reasonably revolutionize the entire research industry.
This was pretty amazing: I got access to the 1 million token Gemini Pro, and fed in the 20 papers and books that made up my academic work prior to 2022, over 1000+ pages of PDFs
It was able to extract direct quotes & find themes across all of them with only quite minor errors...
— Ethan Mollick (@emollick)
2:56 AM • Feb 20, 2024
An astute Redditor pointed out that “Whether Gemini 1.5 can do that isn't even that relevant; what is relevant is that in a not so distant future we will have AI capable of reading every book and research paper ever written, remember everything with perfect recollection and combine the information in new creative ways.”
Now imagine a world where we have Gemini 1.5 and more advanced models running on Groq, providing instant outputs via mixed or virtual reality. This is not a future pipe dream but something we will likely see in the next few months.
I do not think it is outlandish to say that the Gemini 1.5 update is as big of a release as we have had in many months.
While the beta is still closed to researchers and similar power users, there are already prompting guides that will work well for this model. Becoming a Gemini expert and understanding what data, what prompts, and what settings apply for which use cases could lead to a very lucrative career down the line.
Imagine a world where you can start feeding Gemini large datasets and documents from seemingly unrelated fields, and putting the model to work at finding deep connections that would be generally beyond the human capacity to discover.
We are headed toward a whole new world. Some would say a Renaissance.
Artists’ Gallery
ChatGPT Can Help You Build Worlds
Tools
Must-have tools for every Renaissance creator to add to their toolkit:
Kaliber: Turn text, videos, photos, and music into stunning videos with our advanced AI generation engine.
MyShell: Create your own apps with No Code
Gemini Code Editor is now live
Rodin 3D is getting a lot better every version
AlphaCodium: Advancing the Future of Code Generation
DeftGPT is an AI-powered tool that offers a versatile platform for content creation and idea generation
V0 Report: Business reports in 1 click just by entering a URL
Otio: Your AI research and writing partner
TheyDo: Pinpoint, prioritize and solve customer pain points, fast.
Deep Tech
The newest and coolest in the research world that you need to know about:
LLM tokenizer lecture from OpenAI founder Andrej Karpathy
TikTok just announced their new video model called Boximator
Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research
Gemini 1.5 could completely revolutionize the research industry
Meta released a data set of recordings from smart glasses
Groq vs NVIDIA chip architecture
Gemini 1.5 Model Prompting guide
Video explainer on V-JEPA
Yann Lecun explains why a house cat is smarter than the best AI
Intel extensions model released to help LLMs perform better
RLVF: Learning from Verbal Feedback without Overgeneralization
Closing Thought
The AI / VR moment is coming soon. When you can run models like Gemini 1.5 with large context windows on upgraded hardware like Groq, you can essentially create new worlds in real time. It’s coming soon, and I’m not even a VR bull.
Work With Us!
The AI Renaissance is coming and we are building the best community of the people making it happen.
Contact us to sponsor your product or brand and reach the exact audience for your needs across our newsletter and podcast network.