NeurIPS 2023 recap

Recapping the most compelling ideas and emerging research from our time at NeurIPS 2023, the world's premier AI conference.

  • Cutting-Edge Research: NeurIPS is one of the premier conferences in machine learning and artificial intelligence, attracting top researchers worldwide. It’s a hub for the latest advancements in NLP, offering insights into state-of-the-art models, techniques, and methodologies.

  • Networking Opportunities: NeurIPS brings together experts, researchers, and practitioners in the field of NLP. Attending the conference provides us with ample opportunities to network with peers, exchange ideas, and establish connections that could lead to collaborations or future learning opportunities.

  • Learning from Keynotes and Workshops: NeurIPS features keynote presentations by leading figures in AI and NLP, as well as workshops focusing on specialized topics within the field. These sessions offer valuable insights, perspectives, and practical knowledge that can enhance our understanding and expertise in NLP.

  • Paper Presentations and Poster Sessions: NeurIPS showcases a wide range of research papers and poster presentations covering various aspects of NLP, including novel algorithms, applications, and theoretical advancements. Attending these sessions allows us to stay updated on the latest developments and gain inspiration for our work.

  • Industry Trends and Applications: NeurIPS isn’t solely focused on academic research; it also highlights industry trends and applications of NLP.

Specific highlights from Arijit

I will mostly talk about the panel discussion and the keynotes.

  1. Training Language models is easy. With proliferation of well-curated datasets and infrastructure like HuggingFace, even an undergraduate can train a LM on a GPU or even multiple GPUs hosted on a single machine. However the main challenge is to scale this to more GPUs across machines or even data centers spread across the world.

  2. With the scarcity of data for model training, researchers are delving into different methodologies. One avenue involves training multimodal models using text, video, images, and audio together, hoping that skills acquired from diverse modalities can transfer to text. There was also a discussion about using synthetic data to address this data challenge.

  3. There was a discussion whether open foundation models are potentially harmful for AI safety, as they can be exploited by malicious actors. However, Dr. Liang argues that open models also contribute positively to safety. He argues that by being accessible, they provide more researchers with opportunities to conduct AI safety research and to review the models for potential vulnerabilities.

  4. Moreover, there was a discussion about using autoregressive models for image/video generation. Specifically, this was done in the Gemini project. There have also been explorations into using diffusion models for text generation, but these have not yet been proven effective.

  5. Annotating data requires a notably higher level of expertise in the annotation field than it did five years ago. With the anticipated performance boost of AI assistants in the future, gathering valuable feedback data from users may lessen dependence on extensive data from annotators.

Specific highlights from Shirley

As an AI product designer, my favorite parts of the conference involve real-world AI and human-AI system design applications. Here are some of my favorite learnings:

  • Beginner level: Complimentary (AI that: knows when it doesn’t know, can defer to humans)

  • Intermediate level: Co-operative (AI that: knows what it doesn’t know and can ask for more information)

  • Advanced level: Fully interactive (AI that: knows when it doesn’t know, can ask for more information, can interact with humans to make the best decision)

  • You can work with 3rd party vendors

  • You can work with IP restrictions

  • You can also work with regulatory requirements

  • Simple to implement, recalibrate under distribution shift

Previous
Previous

Analyzing complex documents just got easier

Next
Next

Learnings from the lab: Querying S&P Global’s tabular data using LLMs