Ilya Sutskever, OpenAI co-founder, launches Safe Superintelligence (SSI) after leaving the company

  • Ilya Sutskever, co-founder of OpenAI, launches Safe Superintelligence (SSI) after departing from OpenAI.
  • SSI focuses solely on advancing safe superintelligence, aiming for clarity and unwavering commitment.
  • Key team members include Daniel Gross from Apple and Daniel Levy from OpenAI.
  • Headquarters based in Palo Alto, California, with a presence in Tel Aviv.
  • Sutskever’s departure signals shifts in AI research and governance strategies.

Main AI News:

OpenAI co-founder Ilya Sutskever has embarked on a new chapter with the launch of his latest AI endeavor, Safe Superintelligence (SSI). This move follows his departure from OpenAI, where he held the role of chief scientist and co-led the Superalignment team alongside Jan Leike, who also recently left to join rival AI firm Anthropic.

At the core of Sutskever’s new venture is a steadfast commitment to advancing safe superintelligence. In a statement posted on X, Sutskever announced, “I am starting a new company. We will pursue safe superintelligence in a straight shot, with one focus, one goal, and one product.” This singular focus is intended to eliminate distractions typically associated with management overhead and short-term commercial pressures.

SSI’s vision and strategic roadmap are entirely dedicated to enhancing AI safety and security. The startup’s approach ensures that every aspect of its development aligns with the overarching goal of safe and responsible AI advancement. The founding team includes Daniel Gross, renowned for his leadership in AI and search at Apple, and Daniel Levy, a former colleague from OpenAI.

Headquartered in Palo Alto, California, with an additional office in Tel Aviv, SSI aims to be a pivotal player in shaping the future of AI technology. Sutskever’s decision to launch SSI comes amidst significant changes within OpenAI, including boardroom disputes over AI governance, which saw Sutskever and others advocating for robust guardrails in AI development.

Despite past challenges, including controversies over leadership and strategic direction, Sutskever remains dedicated to leveraging AI for positive impact while ensuring ethical considerations and safety protocols are prioritized. SSI’s establishment underscores Sutskever’s ongoing influence in the AI community and his commitment to advancing the field responsibly.


Ilya Sutskever’s launch of Safe Superintelligence marks a significant development in the AI market. By focusing squarely on AI safety and ethical considerations, Sutskever not only addresses critical gaps in current AI governance but also sets a new standard for responsible innovation. This move is likely to influence industry practices, encouraging a deeper integration of safety protocols and ethical frameworks into AI development strategies globally.