Explaining computerized English testing in plain English

app Languages
a pair of hands typing at a laptop

Research has shown that automated scoring can give more reliable and objective results than human examiners when evaluating a person’s mastery of English. This is because an automated scoring system is impartial, unlike humans, who can be influenced by irrelevant factors such as a test taker’s appearance or body language. Additionally, automated scoring treats regional accents equally, unlike human examiners who may favor accents they are more familiar with. Automated scoring also allows individual features of a spoken or written test question response to be analyzed independent of one another, so that a weakness in one area of language does not affect the scoring of other areas.

was created in response to the demand for a more accurate, objective, secure and relevant test of English. Our automated scoring system is a central feature of the test, and vital to ensuring the delivery of accurate, objective and relevant results – no matter who the test-taker is or where the test is taken.

Development and validation of the scoring system to ensure accuracy

PTE Academic’s automated scoring system was developed after extensive research and field testing. A prototype test was developed and administered to a sample of more than 10,000 test takers from 158 different countries, speaking 126 different native languages. This data was collected and used to train the automated scoring engines for both the written and spoken PTE Academic items.

To do this, multiple trained human markers assess each answer. Those results are used as the training material for machine learning algorithms, similar to those used by systems like Google Search or Apple’s Siri. The model makes initial guesses as to the scores each response should get, then consults the actual scores to see well how it did, adjusts itself in a few directions, then goes through the training set over and over again, adjusting and improving until it arrives at a maximally correct solution – a solution that ideally gets very close to predicting the set of human ratings.

Once trained up and performing at a high level, this model is used as a marking algorithm, able to score new responses just like human markers would. Correlations between scores given by this system and trained human markers are quite high. The standard error of measurement between app’s system and a human rater is less than that between one human rater and another – in other words, the machine scores are more accurate than those given by a pair of human raters, because much of the bias and unreliability has been squeezed out of them. In general, you can think of a machine scoring system as one that takes the best stuff out of human ratings, then acts like an idealized human marker.

app conducts scoring validation studies to ensure that the machine scores are consistently comparable to ratings given by skilled human raters. Here, a new set of test-taker responses (never seen by the machine) are scored by both human raters and by the automated scoring system. Research has demonstrated that the automated scoring technology underlying PTE Academic produces scores comparable to those obtained from careful human experts. This means that the automated system “acts” like a human rater when assessing test takers’ language skills, but does so with a machine's precision, consistency and objectivity.

Scoring speaking responses with app’s Ordinate technology

The spoken portion of PTE Academic is automatically scored using app’s Ordinate technology. Ordinate technology results from years of research in speech recognition, statistical modeling, linguistics and testing theory. The technology uses a proprietary speech processing system that is specifically designed to analyze and automatically score speech from fluent and second-language English speakers. The Ordinate scoring system collects hundreds of pieces of information from the test takers’ spoken responses in addition to just the words, such as pace, timing and rhythm, as well as the power of their voice, emphasis, intonation and accuracy of pronunciation. It is trained to recognize even somewhat mispronounced words, and quickly evaluates the content, relevance and coherence of the response. In particular, the meaning of the spoken response is evaluated, making it possible for these models to assess whether or not what was said deserves a high score.

Scoring writing responses with Intelligent Essay Assessor™ (IEA)

The written portion of PTE Academic is scored using the Intelligent Essay Assessor™ (IEA), an automated scoring tool powered by app’s state-of-the-art Knowledge Analysis Technologies™ (KAT) engine. Based on more than 20 years of research and development, the KAT engine automatically evaluates the meaning of text, such as an essay written by a student in response to a particular prompt. The KAT engine evaluates writing as accurately as skilled human raters using a proprietary application of the mathematical approach known as Latent Semantic Analysis (LSA). LSA evaluates the meaning of language by analyzing large bodies of relevant text and their meanings. Therefore, using LSA, the KAT engine can understand the meaning of text much like a human.

What aspects of English does PTE Academic assess?

Written scoring

Spoken scoring

  • Word choice
  • Grammar and mechanics
  • Progression of ideas
  • Organization
  • Style, tone
  • Paragraph structure
  • Development, coherence
  • Point of view
  • Task completion
  • Sentence mastery
  • Content
  • Vocabulary
  • Accuracy
  • Pronunciation
  • Intonation
  • Fluency
  • Expressiveness
  • Pragmatics

More blogs from app

  • Grammar 101: insider tips and tricks to instantly improve your writing (part 4)

    By
    Reading time: 7 minutes

    Punctuation makes your writing easier to read and understand, but it can be tricky to master. As an editor and proofreader, I often notice people confusing semi-colons and colons, so we'll explore the difference between them. And because both are often used in lists, we'll also look at the humble comma – and its sometimes-controversional cousin, the Oxford comma.

    Semi-colons and colons both connect phrases in a sentence but are used in different situations.

    Understanding colons

    Colons introduce important information and explanations. They're often used before lists as a replacement for phrases like "they are" and "which is":

    • He offered me a choice of drinks: tea, coffee or hot chocolate.
    • I packed the essentials in my bag: water, pens and a magazine.
    • She speaks three languages: English, French and Portuguese.

    You can also think of a colon as a spotlight, with the phrase that comes after the colon explaining or expanding what came before it.

    • In 1903, travel was changed forever by an important event: Orville and Wilbur Wright's first successful flight.
    • He loves visiting the animals at the farm: cows are his favourite.
    • There is one rule I live by: I treat others as I wish to be treated.

    The secrets of semi-colons

    A semi-colon links two ideas that are closely related and that would be two complete sentences if you used a period instead. They give a softer transition than a period would, and they're often used instead of conjunctions like "and", "but" and "because":

    • I love eating pizza; my sister loves eating burgers.
    • I wanted to go for a swim; I couldn't find my goggles.
    • I'm going to the park; the weather is beautiful.

    Semi-colons also seperate items in long lists to make life easier for the reader and stop a sentence becoming a sea of apostrophes. For example:

    • I've got my shopping list ready: peppers, carrots and oranges from the market; toothpaste, shampoo and pain relief from the drugstore; and a newspaper, snack and drink from the newsstand.

    Standard comma or Oxford comma?

    An Oxford comma goes before "and" or "or" at the end of a list. The first example has an Oxford comma, the second doesn't.

    • Please bring me a sandwich made with cheese, lettuce, and tomato.
    • Please bring me a sandwich made with cheese, lettuce and tomato.

    American English generally favors the Oxford comma, British English typically omits it, unless needed for clarity. Compare:

    • I love my parents, Taylor Swift and Keanu Reeves.
    • I love my parents, Taylor Swift, and Keanu Reeves.

    As with many areas of punctuation, whether you choose to use the Oxford comma is a matter of personal preference. However, the most important thing is to be consistent in your usage.

  • Precision teaching with AI: Aligning GSE objectives with generative AI for targeted materials

    By
    Reading time: 4 minutes

    English teachers today face increasing demands: create engaging content, differentiate instruction and address diverse learner needs – all within a limited time. The rise of Generative AI, like ChatGPT, offers a promising solution. But without proper guidance, AI-generated content can lack educational value. This blog post introduces a practical, research-informed approach to using AI tools aligned with the Global Scale of English (GSE). You will learn how this framework helps educators design accurate, personalized and level-appropriate English teaching materials quickly and confidently.

    Why GSE and AI are a game-changing combination for ELT

    The Global Scale of English (GSE) is a CEFR-aligned framework developed by app, offering detailed "can-do" learning objectives. It includes nearly 4,000 descriptors across speaking, listening, reading and writing skills, offering more precision than traditional level labels like A2 or B1. At the same time, Generative AI tools such as ChatGPT can generate entire lessons, tasks and assessments in seconds. The challenge lies in ensuring this content is aligned with clear pedagogical outcomes.

    Pairing AI’s creative speed with the GSE’s structured outcomes offers a scalable way to meet learner needs without compromising instructional quality.

    Unlocking measurable, differentiated and efficient teaching with GSE and AI

    The GSE makes objectives measurable

    Unlike generic teaching goals, GSE objectives are specific and measurable. For example, a B1-level learner objective might state:

    “Can identify a simple chronological sequence in a recorded narrative or dialogue.” (GSE 43)
    This clarity helps teachers define outcomes and ensure each AI-generated task targets an actual language skill, not just generic content.

    Generative AI enhances productivity

    Teachers using Generative AI can create draft lesson materials in minutes. By inputing a structured prompt such as:

    “Create a B1 reading activity that helps learners summarize the main points of a short article.”
    ChatGPT can instantly generate content that meets the learning goal. When guided by the GSE, AI becomes a collaborative assistant as well as a time-saver.

    The GSE + AI combination supports differentiation

    Because the GSE includes descriptors across a wide proficiency range (from pre-A1 to C2), teachers can tailor AI-generated content to meet the exact needs of their students. Mixed-level classrooms or tutoring contexts benefit especially from this, as teachers can create multiple versions of a task with consistent scaffolding.

    Practical tips

    • Use the GSE Teacher Toolkit to select objectives based on skill, level or function.
    • When prompting ChatGPT, include the GSE descriptor in your input for more precise results.
    • Always review and adapt the AI output to match your learners’ context, culture and curriculum.
    • Create a prompt library mapped to GSE codes to save time in future planning.

    A step-by-step example of the GSE and AI in action

    Here is a typical application of the workflow:

    1. A teacher selects a GSE objective, such as:
      “Can write a basic formal email/letter requesting information.” (GSE 46).
    2. Within seconds, a sample formal email, accompanied by a short reading comprehension task and a vocabulary activity, is generated.
    3. The reading task serves as a model to help learners analyze the structure, tone, and key language features of a well-written email before attempting their own.
    4. The teacher then reviews and refines the output for clarity, appropriateness, and context relevance.

    This process supports targeted teaching while significantly reducing preparation time.

    Overcoming challenges: Ensuring quality and relevance

    Challenge: AI outputs may lack cultural context, level appropriateness or instructional clarity.
    Solution: Always pair AI with professional judgment. Use the GSE to check that skills match the intended outcome, and adjust the complexity of the language as needed.

    Challenge: Teachers may be unfamiliar with how to write effective AI prompts.
    Solution: Start simple with templates like:

    “Create a [skill] activity at [level] that supports this GSE objective: [insert objective].”

    Challenge: Risk of over-relying on AI for instruction.
    Solution: Use AI as a starting point, not the final product. Combine AI-generated content with classroom interaction, feedback and your own creativity.

    Teaching tools that make this easier

    • : for exploring and selecting level-appropriate learning objectives
    • : for generating customizable teaching content
    • GSE Smart Lesson Generator: an AI-powered lesson creation tool developed by app that uses the GSE framework to automatically generate high-quality activities and lesson plans
    • Google Docs or Word: for editing and organizing your materials before class

    Confidently transforming English teaching

    Combining Generative AI with the Global Scale of English allows teachers to design materials that are both fast and focused. The GSE provides the structure; AI provides the speed. Together, they offer a sustainable solution for personalized English instruction that respects both learner needs and instructional quality.

  • woman writing in notepad while looking at laptop computer and smiling

    Grammar 101: insider tips and tricks to instantly improve your writing (part 3)

    By
    Reading time: 7 minutes

    Many people can't tell the difference between the hyphen (-), the en-dash (–), and the em-dash (—). They may look similar but they can all help ensure that your writing looks professional and is easy to read. As an overview:

    • Hyphens improve clarity: there is a big difference between "a man-eating shark" and "a man eating shark".
    • En-dashes and em-dashes share a lot of the same functionality – including allowing for explanations and examples to be shared, and separating clauses – however, they are not interchangeable and their use is often down to personal preference.

    Let's explore what these three different dashes do and how they could improve your writing.