S
Awaira Score
80
Out of 100
Valuation
N/A
Post-money
Total Raised
$1B
All rounds
Awaira Score
80/100
Founded
2024
50-200 employees
What They Build
March 2026Safe Superintelligence (SSI) is an AI safety research organization focused exclusively on building safe superintelligent AI systems, operating with a deliberate separation between its safety research mission and commercial product pressure. The company is structured to pursue the long-term goal of superintelligence with safety as the primary engineering and research objective rather than a seconda…
Is this your company? Claim it →Company Info
StageSeries A
Employees50-200
Country🇺🇸 United States
Share
Loading sentiment...
Funding Rounds
Series A · No public funding round data available yet.
Founded Same Year (2024)
More from United States
🇺🇸 View all AI companies in United States →Frequently Asked Questions
What is Safe Superintelligence's valuation?▾
Safe Superintelligence's valuation is not publicly disclosed.
Who invested in Safe Superintelligence?▾
Investor information for Safe Superintelligence is not publicly available at this time.
When did Safe Superintelligence last raise funding?▾
No public funding round data is currently available for Safe Superintelligence.
How many employees does Safe Superintelligence have?▾
Safe Superintelligence has approximately 50-200 employees.
What does Safe Superintelligence do?▾
Safe Superintelligence (SSI) is an AI safety research organization focused exclusively on building safe superintelligent AI systems, operating with a deliberate separation between its safety research mission and commercial product pressure. The company is structured to pursue the long-term goal of superintelligence with safety as the primary engineering and research objective rather than a secondary consideration added to a commercial AI product.\n\nThe company raised approximately 1 billion USD in one of the highest-profile AI funding rounds of 2024, backed by investors including Andreessen Horowitz and Sequoia Capital. SSI was co-founded by Ilya Sutskever, the former OpenAI Chief Scientist and co-inventor of foundational deep learning techniques, and Daniel Gross, giving the organization exceptional technical credibility in the AI safety research community.\n\nSafe superintelligence research represents the frontier of the AI capability development race, where a small number of organizations are competing to determine whether advanced AI systems can be built with verifiable safety guarantees. SSI unusual structure, which deliberately avoids near-term product commitments that could compromise safety research, positions it as one of the most watched AI organizations in the world regardless of near-term commercial output.