AI in Science
AI in Science
Introduction
Many scientists now use AI tools. Some people like these tools. Other people worry about the truth in science.
Main Body
More scientists use AI every year. Some use it to fix their English. Other scientists do not use AI. They want to learn and be honest. AI makes mistakes. It writes things that are not true. It also uses a lot of water and electricity. This is bad for the planet. Many people send AI papers to science journals. It is hard to know if a human or a computer wrote the text. This is a big problem for science.
Conclusion
Scientists disagree about AI. They want to work fast, but they also want the work to be correct.
Learning
π‘ The 'Some vs. Other' Pattern
When we talk about two different groups of people, we use this simple pair:
- Some (Group A) Some people like AI.
- Other (Group B) Other people worry.
π οΈ Word Swap: Simple Opposites
Look at how the text shows a conflict using basic words:
| Positive/Fast | Negative/Slow |
|---|---|
| Like Worry | Work fast Be correct |
| Fix English Be honest | Human Computer |
π Quick Tip: 'Bad for...'
To describe a negative result, use this structure:
[Something] is bad for [Something else].
- Example: Using AI is bad for the planet.
Vocabulary Learning
The Rise of Generative AI in Scientific Research and the Risks to Institutions
Introduction
The use of large language models (LLMs) in academic work has caused a tension between the desire for faster work and the need to maintain scientific honesty.
Main Body
There is currently a big difference in how researchers use generative AI. For example, an Elsevier survey showed that usage rose from 37% to 58% in one year. However, a Nature survey suggests that most researchers only use AI to improve their language, while very few use it to write the main text. Some scholars avoid these tools entirely because they want to develop their own thinking skills and avoid ethical problems regarding where the data comes from. Institutional stability is also threatened by 'hallucinations,' which are factual errors made by AI. In chemistry and conservation science, AI often creates fake molecular structures and wrong citations. Consequently, humans must spend a lot of time checking the work, which removes the time-saving benefits of the technology. Furthermore, these tools have a high environmental cost; it is estimated that by 2025, they will produce millions of tonnes of CO2 and use vast amounts of water, which contradicts the goals of climate research. Finally, there is a growing conflict over how to detect AI-written content. Submissions to 'Organization Science' increased by 42% after ChatGPT was released, and many of these papers contained over 70% AI-generated text. Similarly, AI-generated reviews in computer science rose from 7% in 2023 to 43% in 2025. Because current detection tools cannot always tell the difference between AI editing and full AI generation, fake data may enter the official scientific record.
Conclusion
The scientific community remains divided as it tries to balance faster research production with the need for high quality and ethical standards.
Learning
The 'Cause and Effect' Upgrade
An A2 student usually says: "AI makes mistakes. So, humans check the work."
A B2 student uses connecting adverbs to show a logical flow. This makes your writing sound academic and professional.
The Power Word: Consequently
In the text, we see: "AI often creates fake molecular structures... Consequently, humans must spend a lot of time checking the work."
Why this is a B2 move:
Instead of using "so" (which is very basic), Consequently signals that the second sentence is a direct, logical result of the first. It transforms a simple observation into a formal argument.
Expanding Your Logical Toolkit
To reach B2, you need to vary how you connect ideas. Look at these shifts based on the article's themes:
| A2 Style (Basic) | B2 Style (Advanced) | Transition Word |
|---|---|---|
| AI is fast, but it is risky. | AI offers speed; however, it introduces significant risks. | However |
| AI uses water. It also produces CO2. | AI uses vast amounts of water; furthermore, it produces millions of tonnes of CO2. | Furthermore |
| AI creates fake data. This is why it's dangerous. | AI creates fake data; therefore, the scientific record is threatened. | Therefore |
Quick Guide to Usage
- However: Use this when you want to pivot to an opposite idea.
- Furthermore: Use this when you are adding a "second layer" of evidence to your point.
- Consequently / Therefore: Use these when the second fact is a result of the first.
Pro Tip: Notice that these words are often followed by a comma (,) when they start a new sentence. This is a key marker of B2-level punctuation.
Vocabulary Learning
The Proliferation of Generative Artificial Intelligence within Scientific Research and Associated Institutional Risks
Introduction
The integration of large language models (LLMs) into academic workflows has precipitated a systemic tension between perceived operational efficiency and the preservation of scientific integrity.
Main Body
The current academic landscape is characterized by a significant divergence in the adoption of generative AI (genAI). Quantitative data indicate a rising trend in utilization; an Elsevier survey reported an increase in researcher usage from 37% to 58% between the previous and current year. Conversely, a Nature survey suggests a more cautious consensus, where a vast majority of respondents accept AI for linguistic refinement, yet a minority utilize it for primary text generation. This dichotomy is further reflected in individual practitioner behaviors, where some scholars purposefully abstain from these tools to ensure the development of foundational cognitive skills and to avoid the ethical complications associated with data provenance. Institutional stability is further challenged by the emergence of 'hallucinations' and factual inaccuracies. Evidence from the fields of chemistry and conservation science indicates that genAI frequently produces nonsensical molecular structures and erroneous citations, necessitating rigorous human verification that often negates the intended efficiency gains. Furthermore, the environmental externalities of these technologies are substantial; projections for 2025 estimate a global carbon footprint between 32.6 and 79.7 million tonnes of CO2, alongside significant water consumption, which some researchers argue is antithetical to the objectives of climate-centric research. Concurrent with these internal debates is an escalating conflict regarding the detection of AI-generated content. Analysis of manuscripts submitted to 'Organization Science' revealed a 42% increase in submissions following the release of ChatGPT, with a corresponding rise in papers containing over 70% AI-generated text. Similarly, data from arXiv indicates that AI-generated review preprints in computer science rose from 7% in 2023 to 43% in 2025. The inability of current detection tools to consistently distinguish between AI-assisted editing and wholesale generation creates a vulnerability in the peer-review process, potentially permitting the infiltration of fabricated data into the scientific canon.
Conclusion
The scientific community remains divided as it attempts to balance the acceleration of research output with the necessity of maintaining rigorous quality control and ethical standards.
Learning
The Architecture of 'Nominal Density' and Conceptual Compression
To bridge the gap from B2 to C2, a student must move beyond describing events and begin conceptualizing them through Nominalization. The provided text is a masterclass in this; it does not simply say "AI is being used more, which causes problems," but rather utilizes dense noun phrases to encapsulate complex causal relationships.
β‘ The C2 Pivot: From Action to Entity
Observe the transition from a standard academic sentence to the text's high-density construction:
- B2 Approach: The integration of LLMs into academic work has caused a tension between how efficient things seem and how we keep science honest.
- C2 Execution: The integration of large language models (LLMs) into academic workflows has precipitated a systemic tension between perceived operational efficiency and the preservation of scientific integrity.
Analysis: The author replaces verbs (caused, keep) with nouns (precipitated, preservation). This transforms a sequence of actions into a conceptual landscape. In C2 English, nouns act as "containers" for complex ideas, allowing the writer to manipulate entire theories as single objects.
π Deconstructing the 'Lexical Clusters'
Look at the phrase: Environmental externalities of these technologies.
- Externalities is a high-level economic term. By pairing it with Environmental, the author avoids a lengthy explanation of "the side effects of pollution caused by companies."
- This is Conceptual Compression. A C2 writer uses precise, multi-disciplinary terminology to signal expertise and maintain a formal, detached register.
π οΈ Application: The 'Sovereign Noun' Technique
To emulate this, focus on the Sovereign Nounβthe headword that governs the entire clause.
Example from text:
...the infiltration of fabricated data into the scientific canon.
Breakdown:
- The Head: Infiltration (The core concept: something entering where it shouldn't).
- The Modifier: Fabricated data (The object being infiltrated).
- The Destination: Scientific canon (The prestige/authority being threatened).
By centering the sentence on Infiltration rather than saying "AI might let fake data get into journals," the writer elevates the discourse from a simple warning to a systemic critique.