Using Large Language Models for Qualitative Analysis can Introduce Serious Bias

Large Language Models (LLMs) are quickly becoming ubiquitous, but the implications for social science research are not yet well understood. This paper asks whether LLMs can help us analyse large-N qualitative data from open-ended interviews, with an application to transcripts of interviews with Rohingya refugees in Cox's Bazaar, Bangladesh. We find that a great deal of caution is needed in using LLMs to annotate text as there is a risk of introducing biases that can lead to misleading inferences. We here mean bias in the technical sense, that the errors that LLMs make in annotating interview transcripts are not random with respect to the characteristics of the interview subjects. Training simpler supervised models on high-quality human annotations with flexible coding leads to less measurement error and bias than LLM annotations. Therefore, given that some high quality annotations are necessary in order to asses whether an LLM introduces bias, we argue that it is probably preferable to train a bespoke model on these annotations than it is to use an LLM for annotation.


Introduction
Large Language Models (LLMs) are increasingly being used in social science research to, among other things, analyze and annotate text data (Gilardi et al., 2023). As LLMs become more accessible and popular we can expect that there will be a temptation to use them to analyze open-ended interview data such as those used by qualitative researchers (Small and Calarco, 2022) who follow a descriptive/interpretative analytical approach (Elliot and Timulak, 2005). This relies on careful, nuanced, coding conducted by trained social scientists (Detering and Waters, 2018). Qualitative analysis of this kind lies at the core of fields like anthropology and sociology but is now being increasingly employed in more quantitative fields such as economics because of the ability of Natural Language Processing (NLP) methods to analyze open-ended data at scale (Rao, 2023).
Data generated from open-ended, in-depth, interviews is potentially very different from the benchmark datasets often used in the NLP literature to validate modelling approaches such as English language tweets and news, or product reviews. This is because qualitative research is often conducted in a manner where the specific context matters in interpreting the data, and analyzed with codes that are "flexibly" developed that can be quite nuanced and complex. Our example application falls into this category -we have interviews on a very specific topic with a very specific population who are not well represented in the training data that LLMs are trained on (or in the data used in the NLP literature more broadly).
We find that in such a context, using LLMs to annotate text is potentially dangerous. We test three different LLMs (ChatGPT and two versions of Meta's Llama 2) and find that the prediction errors they make in annotation are not random with respect to the characteristics on the interview subject. This can lead to misleading conclusions in later analysis, as we shown in Figure 5. Statistical analysis based on LLM annotations can lead to estimated effects that are very different from those based on human expert annotations. It is therefore crucial to have some high quality expert annotations, even if it is just to assess whether the LLM is introducing bias or not. Given that some high quality annotations are needed to assess whether the LLM introduces bias, we argue that it is preferable to train a bespoke model on these annotations than it is to use an LLM.
We show that iQual, a method we developed with others (Ashwin et al., 2022) to analyze large-N qualitative data by training supervised models on small human annotated samples, not only performs better than LLMs in terms of out-of-sample prediction accuracy but also introduces much less bias. LLMs can possibly assist this process by generating larger training sets (i.e. data augmentation, as proposed by Dai et al. (2023)) but we only find evidence of marginal benefits in a few cases. This suggests a potential way in which to reconcile the nuance and "reflexive" qualities of interpretative qualitative analysis with large representative samples. Crucially, we see LLMs and other NLP methods as assisting and extending traditional qualitative analysis, not replacing it. In order to create a coding tree that captures important and interesting variation across documents in a nuanced and contextaware manner, there is no substitute for a careful reading on at least a subset of those documents.
Our application is based on open-ended interviews with Rohingya refugees and their Bangladeshi hosts in Cox's Bazaar, Bangladesh. These interviews focused on subjects aspirations and ambitions for their children (Callard, 2018) as well as their capacity to achieve those goals, i.e. their navigational capacity (Appadurai, 2004). They are analysed in detail in Ashwin et al. (2022), so we will not discuss the detail of data collection or related social science literature here. The substance of these interviews is not critical to the methodological contribution of this paper, but it is important to note that while "ambition" can be captured well by structured questions that yield quantitative data, aspirations and navigational capacity are subtle and complex concepts not easily defined are captured in structured surveys. It is precisely when dealing with these sorts of concepts that open-ended interviews and interpretative qualitative analysis is valuable. The complexity and nuance of the concepts may play a role in explaining the poor performance of LLMs in annotating interviews compared with other studies where the annotation tasks were substantially more straightforward, e.g. Mellon et al. (2022).
Previous work has suggested that LLMs might outperform crowd-sourced human annotations (Gilardi et al., 2023), or even that a substantial proportion of workers on crowd-sourcing platform may be using LLMs in completing tasks (Veselovsky et al., 2023). Our results do not contradict these as for many annotation tasks LLMs may indeed perform very well and save researchers the expense and complication of crowd-sourcing. However, our results do suggest that researchers ought to be aware of the possibility of biases introduced by LLM annotation, particularly on data where a nuanced, contextual understanding of the documents is needed; LLMs reflect the data they are trained on and many of the contexts in which qualitative analysis adds value require an understanding of communities and concepts that may not be adequately represented in this training data.
The paper is structured as follows. The remainder of this Section discusses this paper's contribution in the context of related literature. Section 2 then very briefly introduces our dataset of annotated interview transcripts. Section 3 describes our approach to using LLMs for annotation (3.1) and the supervised NLP method introduced by Ashwin et al. (2022) which we refer to as iQual going forward (3.2). Section 4 then describes LLM-based out-of-sample performance in comparison to iQual (4.1) and then shows that LLMs introduce more bias and illustrates this could cause researchers to draw incorrect conclusions (4.2). Section 5 then concludes.

Data and Qualitative Analysis
The interview transcripts, data collection and the qualitative coding process are explained in detail in Ashwin et al. (2022), so we restrict ourselves to a very brief description here. The population we sample are Rohingya refugees based in the Cox's Bazzar camp and local Bangladeshi residents. Along with a standard household survey including questions on demographics and economic conditions, the data include transcripts of 2,407 open-ended interviews with subjects on their aspirations for their eldest child. The interviews were conducted either in Bengali or in Rohingya which was then transcribed into Bengali, but we work with machine translations into English. The interviews take the form of an unstructured to-and-fro of question and answer (QA) pairs the interviewer and the subject. The interviews are on average 12.6 QA pairs long, with the average answer in each QA pair being 13.7 words long.
Based on a close reading of a subset of transcripts, and following a "flexible coding" process (Detering and Waters, 2018), a coding tree was developed including 25 potentially overlapping categories, 19 of which we focus on in this paper. A full description of each code along with examples are shown in Appendix A. Following Callard (2018) the distinction between aspiration and ambition was adapted within the context and nature of "dreams" parents expressed for their children. For example, concrete and measurable dreams for child (e.g wishing a child would become a doctor, teacher, entrepreneur, or specific educational goals) was used as a definition for ambition while intangible, value oriented goals (e.g wishing the child to live with dignity or be a good human being) was classified as aspiration. Aspirations, were divided into "Religious" and "Secular" . Ambition was divided into five major categories -Education (further sub-coded into High, Low, Neutral and Religious), Job Secular, Marriage, Entrepreneurship, Migration, Vocational Training, and No Ambition. While ambition and aspiration came up at any point in an interview, "capacity to aspire" or Navigational capacity was restricted to discussions of what have parents were planning or able to do to fulfill dreams for their children. Navigational Capacity was coded into seven sub-codes -Low and High "Ability", Low and High "Budget", Low and High "Information Awareness", and Reliance on God.
Of our sample of 2,407 interview transcripts, 789 are manually annotated by trained sociologists (co-authors on the Ashwin et al. (2022) paper) according to this coding structure. The annotations are defined at the level of QA pairs, allowing us to represent each annotation as a binary classification problem at the QA level.

Methods
In this Section we first explain how we use LLMs to annotated our interview transcripts. We then briefly describe the iQual method which trains supervised models on our expert human annotations, as well as how we use LLMs for data augmentation in combination with iQual. We test three different LLMs-the closed-source ChatGPT (gpt-3.5-turbo) by OpenAI, as well as two open-source LLMs by Meta, the Llama-2 (13b) and its fine-tuned "chat" variant. (Touvron et al., 2023).The base Llama-2 is pretrained on publicly available online data sources. The chat variant is then fine-tuned on publicly available instruction datasets and over 1 million human annotations. This fine-tuning is designed to make the model align with human preferences using techniques such as Reinforcement Learning with Human Feedback (RLHF). ChatGPT is also fine tuned using RLHFFor all three LLMs, our approach to prompting remains consistent.

Annotation with LLMs
We follow several well-established practices to improve the effectiveness of LLMs in annotating our interview transcripts. We provide a prompt that includes precise directives for the LLM, and employ "few-shot learning" (Brown et al., 2020) as well as "chain of thought" prompting (Wei et al., 2022), as explained below. For each code, we created detailed textual instructions, similar to those one would Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their aspirations concerning their children's future education and careers solely in the context of religion, without implying any monetary incentives. Specifically, your role is to discern whether the parent explicitly expresses a desire for their child to pursue a religious path.
Assign the "Religious Aspirations" label only if the parent explicitly articulates aspirations for their child that include becoming a religious scholar, Hafiz, attending a madrassa, learning Arabic, Quran reading, Islamic covering, regular prayer, working in Islamic banks, or being a Islamic scholar. It is crucial not to assign this label if the child is already engaged in any of the mentioned activities, as current religious activities are not considered future aspirations. If there is no such direct reference to religious aspirations in the parent's statement, assign the "Not Applicable" label to indicate that the topic of religious aspirations was not addressed. Here are some examples: ### Input: Interviewer: What are your aspirations for your son's future? Respondent: I hope that he will become a great Islamic scholar. I want him to learn Arabic and read the Quran regularly. ### Output: Interpretation: The parent explicitly expresses a desire for their child to pursue a path in religious education, specifically mentioning becoming an Islamic scholar, learning Arabic, and reading the Quran regularly. Therefore, this conversation aligns with the 'Religious Aspirations' label. Label: Religious Aspirations ### Input: Interviewer: Do you have any specific dreams for your daughter? Respondent: I want her to excel in her studies and become a successful engineer. ### Output: Interpretation: While the parent does express aspirations for their child's future, these aspirations are not religious in nature. The parent hopes for their child to become an engineer, which doesn't involve any religious activities as per the instruction. Therefore, this conversation should be labeled as 'Not Applicable'. Label: Not Applicable ### Input: Interviewer: What do you envision for your child's future? Respondent: I would like my son to become a Hafiz. It's a great honor in our religion. ### Output: Interpretation: The parent explicitly expresses a desire for their child to become a Hafiz, which is a religious aspiration as per the instruction. Therefore, this conversation aligns with the 'Religious Aspiration' label. Label: Religious Aspirations ### Input: Interviewer: What are your hopes for your children's future careers? Respondent: I want them to have successful careers. I hope my daughter becomes a doctor and my son becomes a lawyer. ### Output: Interpretation: While the parent does express aspirations for their children's future, these aspirations do not involve any religious activities or pursuits as per the instruction. Therefore, this conversation should be labeled as 'Not Applicable'. Label: Not Applicable ### Input: Interviewer: What is your dream about your son who is studying in class one? Respondent: Still in Noorani. After reading Noorani, he will read Hafezi. By the grace of Allah, he will teach in a madrasa when he becomes a hafez. That is my dream.
### Output: (ChatGPT) Interpretation: The respondent expresses a clear religious aspiration for their child, which is to become a Hafiz and teach in a madrasa. There is no mention of any monetary incentives, so this response aligns with the criteria for "Religious Aspiration". Label: Religious Aspiration ### Output: (Llama-2-13b) Interpretation: The parent explicitly expresses a desire for their child to pursue a path in religious education, specifically mentioning becoming a Hafiz and teaching in a madrassa. Therefore, this conversation aligns with the "Religious Aspiration" label. Label: Religious Aspiration ### Output: (Llama-2-13b-chat) Interpretation: The parent explicitly expresses a desire for their child to pursue religious studies and become a Hafiz (one who has memorized the Quran). They also mention teaching in a madrasa, which is an Islamic school. Therefore, this conversation aligns with the "Religious Aspiration" label. Label: Religious Aspiration Note: This Figure shows example instructions given to the LLMs for the "Aspirations: Religious" code. These results are given to the LLM as a single prompt, but for clarity we have grouped the instructions into three parts here, as indicated by the colors. The first section (in green) describes the task and gives the definition of the code. The labelled examples (in blue) help the LLM understand the task at hand, often referred to as "few shot learning". For each example, we also provide a reasoning for the annotation (in red) so that the LLM is also asked to explain why it applies a certain label, which is known as "chain of thought" prompting. The second box shows an example of a QA pair to be annotated and the subsequent boxes show the responses of the three LLMs to this prompt.
give to human annotators. These instructions include enough details to ensure that, in principle, the model is fully aware of the specific standards and definitions required for coding transcripts. Each code and each question-answer pair are annotated by the LLM independently. By incorporating both few-shot learning and CoTP, we are in line with best practices and give the LLMs a good chance at annotating accurately. Previous work has shown that using these techniques can help LLM outperform crowd workers in text annotation tasks (Gilardi et al., 2023).
The choice of prompt given to an LLM when giving it a certain task can make a substantial difference to its performance. We give the models a thorough briefing of what each code represents and how to identify its presence in a conversation. This includes the context, certain specific terms or activities that could be indicators, and the need to distinguish between current circumstances and future aspirations, as shown in the green text of Figure 1. These instructions provide a benchmark for the model to understand the coding system and thereby infer the respective codes from the interview transcripts. Instructions for each of the codes are shown in Appendix A.
Few-shot learning and chain of thought prompting (CoTP) are two powerful techniques that can be combined to improve the performance and interpretability of LLMs. Few-shot learning provides examples of a task to the model, which helps guide its behavior and understanding of the task at hand. We provide the LLM with four examples that follow the detailed instructions, as shown by the blue text in Figure 1, to demonstrate correct behavior to the model. These examples are chosen to be instructive of the how the code should be applied and are similar to the examples one would use to explain a code in traditional qualitative analysis.
We also apply chain of thought prompting (CoTP) in these examples to nudge the model to generate an interpretation of the transcript and articulate its line of reasoning before assignment of the final code. It is beneficial in complex tasks where reasoning and interpretation play crucial roles, such as our coding task. The underlying idea is that by having the model outline its thinking process, we can encourage it to reason more deeply and accurately, while also producing outputs that are more interpretable and trustworthy. For our task, we have used both few-shot learning and CoTP by asking the model not only to provide a label for each transcript, but also to give an interpretation explaining why it chose that label.
An example of a full prompt for the 'Religious Aspirations' code are shown in Figure 1, with the instructions and few shot examples for all codes shown in Appendix A.

Training supervised models on interpretative annotations (iQual)
An alternative to using LLMs to annotate large corpora of text documents would be to create high quality annotations on a smaller sub-sample and then training supervised models to predict these annotations on the remainder of the documents. We thus train a separate classifier for each code on a numerical representation of the text at the QA level. As discussed in Appendix B, there are many options for both the classifier we can use here (e.g. random forest, logistic regression, neural networks, SVM), as well as how to represent the text numerically (e.g. tf-idf ngram vectors, sentence embeddings, translations or transliterations). Using k-fold cross-validation we select the best performing model, the text representation and a variety of hyperparameters, so that the approach which performs best in out-of-sample prediction is selected. In each case, we hold out a test set of 200 interviews in order to assess out-of-sample performance. Details about this methodology are provided in Ashwin et al. (2022), and it is implementable in an open source Python package. 1 The crucial intuition though is simply that we use a subset of high quality expert annotations to train a small bespoke model for each code. These models rely only on the annotated training data, unlike the pre-trained LLMs which are trained on huge quantities of text from a huge range of contexts.
Rather than asking LLMs to directly annotate text, another potential use for them is for data augmentation in combination with a supervised model, such as those described above. Data augmentation is a common technique in machine learning to generate more variation in a training set while preserving 1 https://github.com/worldbank/iQual the important signals. For example, when training a model on a labelled dataset of images of animals, one might generate extra variation in the training data by rotating the images by 90 degrees or transforming them into a mirror image of themselves. The idea is to generate more training observations where the noise in the data is different but the signals are the same. A good example of this from the natural language processing literature is back-translation, where text is translated into a different language and then back into the original, so that the exact phrasing and style of the text is different but the meaning is the same (Edunov et al., 2018). Using LLMs for data augmentation has been found to increasing prediction performance in some contexts, so we follow the approach set out in Dai et al. (2023) as an additional experiment here. The example prompts and further details on the augmentation are shown in Appendix A.1.
We thus test two different versions of iQual: first, training supervised models on the human annotations without the use of LLMs, and second, training the model on data augmented by the LLMs to generate more variation in the text while preserving the meaning.

Results
We assess the performance of LLMs in our annotation tasks along two dimensions. Firstly, we assess how accurate of the LLMs predictions relative to our expert human annotations, finding that performance is poor relative to our simpler supervised models. Secondly, and more importantly, we investigate whether the annotations provided by LLMs or iQual introduce bias. We here mean bias in the technical sense that the prediction errors which the models make are not random.

Out-of-sample Performance
Given that the interview transcripts are annotated with a series of binary variables at the QA level, we can assess LLM prediction accuracy with the out-of-sample F1 score for each code. We compare the performance of each LLM to the supervised models trained on annotated data, with and without augmentation. Figure 2 shows the results comparative performance of the different annotation approaches, as measured by the test set F1 score. 2 Given that many of the codes are very sparse, a useful comparison is the F1 score that random guessing would achieve, which is shown as black triangles. All models across all codes do better than random (i.e. have a higher F1 score than that which random guessing would achieve). In all but one case (Capacity: Awareness Information High) ChatGPT is the best performing LLM. However, in all but one case (Capacity: Budget Low) all LLMs perform worse than iQual in terms of these F1 scores. When used for augmentation, ChatGPT does improve performance slightly in some cases, but it slightly worsens performance just as often. If we measure performance in terms of accuracy (i.e. the proportion of observations that are correctly classified) rather than F1 score we get the same results. iQual achieves accuracy of 0.969. In contrast, ChatGPT only achieves 0.909, Llama-2 13B 0.854 abd Llama-2 13B chat 0.851.
These results are of course specific to our context, and a different annotation structure on a different set of text data may lead to different results. However, in our case it is clear that LLMs generate less accurate annotations than training much smaller models on a subset of human annotations does.

Bias
If the annotations generated by LLMs are inaccurate, this is not necessarily a hugely consequential problem. If the mistakes they make are random, with a large enough sample we should still be able to come to correct conclusions. However, if the mistakes are not random, then using LLM annotations can lead to completely incorrect conclusions. In other words, if the LLMs errors are biased, then relying on these annotations could lead researchers to identify relationships in the data that are purely a result of these algorithmic biases rather than reality.
We look at two ways in which the predicted annotations could be biased. Firstly, and most straightforwardly, we show that LLMs over-predict annotations that are very sparse (i.e. there are many more false positives than false negatives). Secondly, we show that in many cases LLM prediction errors are systematically associated with characteristics of the interview subject (e.g. refugee status, gender, education). Figure 3 shows the degree of over-prediction across different annotations. Each model is shown as a separate panel and the bars show the degree of over-prediction as a percentage of all answers. All three LLMs we tested systematically over-predict most of the annotations. This is a problem in itself, as we might be interested in the prevalence of a particular concept, but it is especially problematic if we want to compare the prevalence of different annotations. For example, if we wished to compare the prevalence of secular and religious aspirations in our sample, using the annotations provided by ChatGPT would lead us to very misleading conclusions. While ChatGPT over-predicts both the "Aspirations: Secular" and "Aspirations: Religious" codes, as can be seen from the uppermost two rows of Figure 3, "Aspirations: Secular" is over-predicted much more frequently than "Aspirations: Religious"; in the expert human annotations "Aspirations: Secular" appears around 1.2 times more frequently than "Aspirations: Religious", while in the ChatGPT annotations "Aspirations: Secular" appears around 3 times more frequently than "Aspirations: Religious". Of perhaps even greater concern than over-prediction we find that the LLM's predictions are systematically biased with respect to the interview subjects' characteristics (e.g. refugee status, demographics, education and income). To test whether prediction errors are systematically related to subject characteristics, we regress prediction errors for each model on a range of subject characteristics. We then calculate the F statistic of this regression, which tells us whether there is some statistically significant relationship between the prediction errors and subject characteristics (e.g. a model might over-predict a certain code for men but under-predict for women).  Note: This Figure shows the result of an F-test for a statistical association between the prediction errors of each model with the characteristics of the interview subject. Each model is shown as a separate panel, with each code shown along the vertical axis. The log F statistic of this test is shown along the horizontal axis with the color of the points indicating the statistical significance of the test statistic. The subject characteristics include refugee status; age and sex of eldest child; age, education and sex of interview subject, "refugee"; total number of children; household assets and income; and history of trauma experience. The LLM models display a bias much more frequently that the supervised models. The full results for each regression are shown in Appendix C Figure 4 shows these F statistics that test whether the prediction errors of each annotation approach are systematically related to the interview subjects' characteristics, with the full regression in each case reported in Appendix C. The higher the (log) F statistic is, the stronger the evidence of bias. The color of the points indicates the level of statistical significance and each model is shown as a separate panel. We see that for iQual in the left-most panel, there is evidence of bias in only one of the 19 codes. While we should be cautious in interpreting results with this code, there is not much cause for concern. However, for the LLMs we find strong evidence of bias in many of the codes. This tells us that the prediction errors the LLMs make are not random and conducting analysis on the basis of its predictions is likely to result in misleading interpretations.
The F tests shown in Figure 4 tell us that there is some statistical association between prediction errors and subject characteristics. We can see concrete examples of how this can lead to misleading conclusions in Figure 5. This Figure shows estimated coefficients for regressions of the prevalence of an annotation in an interview on dummy variables for the subjects' refugee status and the gender of their eldest child. So if the coefficient on refugee status is positive than this code appears more in interviews with refugees than in interviews with hosts. Six of the codes are shown as separate panels and the results based on the annotations of the five different models are shown for each coefficient, following the same color scheme as Figures 2 and 3, but with the coefficient based on only the expert annotations shown in black. The differences in the estimates across annotation methods are because the errors they make are not random with respect to refugee status and the gender of the eldest child. We see in Figure 5 that in many cases the coefficient based on the ChatGPT predicted annotations (in pink) is very different from that based on the true human annotations (in black). For example, for "Ambition: Education Low" in the bottom right panel, the coefficient on refugee status is positive and significant using the expert annotations but negative and insignificant on the ChatGPT annotations. The coefficient on whether the eldest child is male meanwhile is negative and insignificant on the human annotations but positive and significant using the ChatGPT annotations. This is just one example, but we can see here how relying on the LLM annotations can lead to potentially dangerous misunderstandings. For example, based on the ChatGPT annotations we might conclude that subjects are more likely to have low educational ambitions for their male children, while in the expert annotations there is no evidence for that. In fact once we increase the sample size using iQual we find a marginally significant effect of the opposite sign.
We can also note here that the coefficients using iQual are not different from those using just the expert annotations, but have much smaller standard errors. Using supervised models to scale up expert human annotations thus increases precision while not introducing bias, as argued in Ashwin et al. (2022). Using ChatGPT to extend the sample size through data augmentation does not appear to introduce additional bias, although it does not have a substantial benefit either. Given that some expert annotations will be necessary in order to identify whether LLM (or crowd sourced) annotations are biased, this suggests that training smaller bespoke models on these annotations may be more reliable than relying on LLMs to annotate large samples.

Discussion
LLMs are trained on a wide range of text and consequently may not be suited for nuanced and context-specific tasks. First, they may introduce systematic biases when used to annotated text. In our example, we see that the errors that LLMs make in annotations (compared to expert human annotations) are not random. Second, LLMs over-predict many of our codes. We can think of this as the LLM bringing the "pre-conceptions" it has learned from its training data to the annotation task. Consequently, LLMs are probably not suited for coding most qualitative data that requires nuanced and contextual analysis. This sort of analysis has traditionally been the province of anthropology and sociology but is increasingly being used by more quantitative fields such as economics and political science.
To analyze large-N qualitative data, such as those obtained from open-ended in-depth interviews, we show that a coding scheme based on a close-reading of transcripts by experts in qualitative analysis with a sub-sample of the full corpus of data is vital for interpretation and analysis. Firstly, high quality annotations are necessary in order to assess whether the LLM is introducing bias in its annotations. Secondly, these high quality annotations can then be used as a training set for smaller bespoke models. These bespoke models may be able to leverage LLMs through data augmentation, but importantly they are trained on context-specific data so researchers have better control, and an overview, of the information that is used. We suspect that these limitations will continue even as LLMs improve, and we encourage researchers using LLMs for annotation tasks to be aware of and check for bias.

System prompt
Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their demonstrated capacity to support their children's education and future aspirations. Specifically, your role is to discern whether the parent explicitly expresses any steps they have taken that go above and beyond their capacity to ensure a better future for their child.
Assign the "High Ability" label only if the parent explicitly articulates measures that they've taken that could be considered extraordinary efforts to secure their child's future. This can include, but is not limited to, investments in education, personal sacrifices, or unique strategies. Keep in mind that financial capacity is not the sole determinant of "High Ability"; parents can still be considered high ability even if they have a limited budget.
If there is no such direct reference to high ability in the parent's statement, assign the "Not Applicable" label to indicate that the topic of high ability was not addressed.

Input:
Interviewer: How are you supporting your child's ambition to become a scientist? Respondent: I've been saving up to send him to a science summer camp every year since he was 10. I also spend time with him doing experiments from his science kit every weekend.

Output:
Interpretation: The parent is making significant efforts to support their child's ambition, including saving money for summer camps and dedicating personal time for science experiments. This demonstrates a high ability to support their child's education and future aspirations. Label: High Ability

Input:
Interviewer: What are you doing to ensure your daughter's future? Respondent: I have been working two jobs to save up for her college education. I also spend my free time helping her with her homework and encouraging her to participate in extracurricular activities that will enhance her college application.

Output:
Interpretation: The parent is making personal sacrifices, such as working multiple jobs and dedicating personal time to support their child's education. This shows a high ability to secure their child's future. Label: High Ability

Input:
Interviewer: How do you plan to help your son achieve his dreams? Respondent: I want him to be happy and successful. I will support him in whatever he chooses to do.

Output:
Interpretation: While the parent expresses a desire to support their child's dreams, they do not specify any extraordinary measures they are taking to ensure a better future for their child. Therefore, this dialogue does not demonstrate high ability. Label: Not Applicable

Input:
Interviewer: What are you doing to help your daughter become a doctor? Respondent: I am doing my best to support her. I always tell her to study hard.

Output:
Interpretation: Although the parent expresses a desire to support their child's ambition, they do not specify any extraordinary measures they are taking to ensure a better future for their child. Therefore, this dialogue does not demonstrate high ability. Label: Not Applicable

Low Ability
System prompt Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their expressed capacity to support their children's education and future aspirations. Specifically, your role is to discern whether the parent explicitly indicates any limitations or constraints in their means to aid their child's future.
Assign the "Low Ability" label only if the parent explicitly states that they have limited or no means (e.g., food, income, resources) to support their child's education and future aspirations.
If there is no such direct reference to low ability in the parent's statement, assign the "Not Applicable" label to indicate that the topic of low ability was not addressed.

Input:
Interviewer: What are your aspirations for your child's education? Respondent: I wish I could give them the best education possible, but I barely make enough to feed them.

Output:
Interpretation: The parent expresses a direct limitation in their ability to support their child's education due to lack of sufficient income. This aligns with the criteria for the "Low Ability" label. Label: Low Ability

Input:
Interviewer: Do you think your children will be able to pursue their dreams? Respondent: They are very ambitious and I'm trying my best to support them.

Output:
Interpretation: Although the parent mentions their effort to support their child's dreams, there is no explicit reference to the limitations or constraints in their means to do so. Hence, the "Low Ability" label is not applicable. Label: Not Applicable

Input:
Interviewer: How are you planning to support your child's dream of becoming a doctor? Respondent: I would love to, but we are struggling to make ends meet. I can't afford to pay for their higher education.

Output:
Interpretation: The parent directly states their inability to support their child's future aspirations due to financial constraints. This meets the criteria for the "Low Ability" label. Label: Low Ability

Input:
Interviewer: Are you able to provide the resources your child needs for his studies? Respondent: Yes, we have managed to provide them with the necessary books and materials.

Output:
Interpretation: The parent confirms their ability to provide resources for their child's education. There is no explicit reference to limitations or constraints in their means to support their child's future. Therefore, the "Low Ability" label is not applicable. Label: Not Applicable

High Awareness
System prompt Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their awareness and understanding of the means to achieve their aspirations for their children's futures. Specifically, your role is to discern whether the parent demonstrates a comprehensive and logical understanding or has the capacity to describe the means of reaching their dreams for their child.
Assign the "High Awareness" label only if the parent demonstrates in their statements a clear understanding and awareness of the processes, resources, and strategies necessary to achieve their aspirations for their child's future. The assignment of this label is inferential, based on your analysis of their comprehension, rather than explicit statements.
If there is no evidence of high awareness in the parent's statement, assign the "Not Applicable" label to indicate that the topic of high awareness was not addressed.

Input:
Interviewer: What do you envision for your son's future? Respondent: I want my son to become a doctor. I've already started saving for his medical school fees and we spend extra time every day working on his science and math homework. I also plan to get him into extracurricular activities that will enhance his application for medical school.

Output:
Interpretation: The respondent demonstrates a clear understanding of the steps necessary to achieve their aspiration for their child's future, which involves him becoming a doctor. They have financial plans in place, focus on relevant academic subjects, and have considered the importance of extracurricular activities. This indicates a high level of awareness about the process. Label: High Awareness

Input:
Interviewer: How are you preparing your child for their future career? Respondent: I believe in my son's capabilities. He's smart and he'll figure it out.

Output:
Interpretation: The respondent does not provide any specific strategies, resources, or processes to prepare their child for their future career. They express faith in their child's abilities but do not demonstrate a high level of awareness about the steps necessary to achieve their career goals. Label: Not Applicable

Input:
Interviewer: What are your thoughts about your daughter's future? Respondent: I want her to be independent and happy.

Output:
Interpretation: While the respondent has a general aspiration for their child's future, they do not provide any specific strategies, resources, or processes to achieve this aspiration.
Therefore, it does not demonstrate a high level of awareness about the steps necessary to achieve this goal. Label: Not Applicable

Input:
Interviewer: What are your dreams for your daughter's future? Respondent: My dream is for her to become a successful engineer. We are ensuring she excels in her math and science classes, and we've enrolled her in a robotics club to give her early exposure to engineering concepts.
We're also exploring scholarship opportunities for engineering programs.

Output:
Interpretation: The respondent demonstrates a clear understanding and awareness of the processes, resources, and strategies necessary to achieve their aspiration for their child's future. They have identified relevant academic subjects, extracurricular activities, and potential financial resources, indicating a high level of awareness about the process. Label: High Awareness

Low Awareness
System prompt Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their awareness and understanding of the means to achieve their aspirations for their children's futures. Specifically, your role is to discern whether the parent demonstrates a lack of knowledge, uncertainty, or confusion regarding how to help their child achieve their aspirations.
Assign the "Low Awareness" label only if the parent demonstrates in their statements a lack of understanding, cluelessness, or uncertainty about the processes, resources, and strategies necessary to achieve their aspirations for their child's future. The assignment of this label is inferential, based on your analysis of their comprehension, rather than explicit statements.
If there is no evidence of low awareness in the parent's statement, assign the "Not Applicable" label to indicate that the topic of low awareness was not addressed.

Input:
Interviewer: What are your plans to ensure your daughter becomes a doctor, which you mentioned is her dream? Respondent: I don't really know. I suppose she will have to study a lot.

Output:
Interpretation: The respondent shows uncertainty about the specific steps and resources required to help their daughter achieve her dream of becoming a doctor. This indicates a lack of awareness about the processes involved, hence the "Low Awareness" label is assigned. Label: Low Awareness

Input:
Interviewer: What are your dreams for your children's future? Respondent: I just want them to be happy and successful. I'm not sure how to make that happen, though.

Output:
Interpretation: Although the respondent expresses aspirations for their children's future, they acknowledge their uncertainty about how to help them achieve success. This suggests a lack of awareness about the necessary strategies and resources, warranting the "Low Awareness" label. Label: Low Awareness

Input:
Interviewer: How are you supporting your son's ambition to be an engineer? Respondent: I'm not sure how to help him. I didn't receive much education myself.

Output:
Interpretation: The respondent admits to not knowing how to assist their son in achieving his ambition of becoming an engineer. This demonstrates a lack of understanding about the necessary steps and resources, thus the "Low Awareness" label is assigned. Label: Low Awareness

Input:
Interviewer: How do you plan to help your children achieve their dreams? Respondent: I believe education is the key. I'm doing my best to save money for their schooling and to provide them with a conducive learning environment at home.

Output:
Interpretation: The respondent demonstrates an understanding of the importance of education in achieving their children's dreams and has a plan to save money for their schooling. This indicates an awareness of the necessary resources and strategies, so the "Low Awareness" label is not applicable. Label: Not Applicable

High Budget
System prompt Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their financial capacity to support their children's education and future aspirations. Specifically, your role is to discern whether the parent explicitly demonstrates having substantial financial resources, either through an ability to save or spend on significant items (e.g., setting up a shop) or services like private tuition or coaching centers.
Assign the "High Budget" label only if the parent explicitly articulates possessing or having access to substantial financial resources that they are able or willing to utilize for their child's education and future aspirations.
If there is no such direct reference to high budget in the parent's statement, assign the "Not Applicable" label to indicate that the topic of high budget was not addressed.
It is important to note that the presence of a parent-child relationship in the conversation, or the discussion of the child's future unrelated to high budget, does not automatically warrant the "High Budget" label. Only discussions that directly and explicitly relate to the parent's financial capacity in supporting their child's future should receive this label.

Input:
Interviewer: What are your plans for your daughter's education? Respondent: I am planning to send her to a private school. I have also arranged for a private tutor to help her with her studies at home.

Output:
Interpretation: The respondent explicitly mentions that they have the financial capacity to send their child to a private school and afford a private tutor, indicating a high budget. Label: High Budget

Input:
Interviewer: How are you helping your children with their studies? Respondent: We are managing somehow. I try to teach them myself when I get time after work.

Output:
Interpretation: The respondent doesn't mention any significant financial investment towards their children's education, indicating that the topic of high budget is not addressed. Label: Not Applicable

Input:
Interviewer: How do you support your son's dream to become a scientist? Respondent: I have enrolled him in a science coaching center and I am also planning to buy him a personal computer for his research work.

Output:
Interpretation: The respondent mentions that they are able to afford enrollment in a coaching center and a personal computer for their child, indicating a high budget. Label: High Budget

Input:
Interviewer: What are your future plans for your child's education? Respondent: I want to send him to a good college, but I'm not sure if I can afford it.

Output:
Interpretation: The respondent mentions a desire to support their child's education but also expresses uncertainty about their financial capacity to do so, indicating that the topic of high budget is not addressed. Label: Not Applicable

Low Budget
System prompt Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their financial capacity to support their children's education and future aspirations. Specifically, your role is to discern whether the parent explicitly indicates a limited financial capacity, directly related to a lack of funds, regardless of their willingness to allocate a budget for their child's education.
Assign the "Low Budget" label only if the parent explicitly articulates a scarcity or lack of financial resources that they have available for their child's education and future aspirations.
If there is no such direct reference to low budget in the parent's statement, assign the "Not Applicable" label to indicate that the topic of low budget was not addressed.

Input:
Interviewer: What are your dreams for your children's future? Respondent: I want them to get a good education and have a successful career, but I'm not sure how we'll afford it. We're barely making ends meet as it is.

Output:
Interpretation: The respondent explicitly expresses a concern about their financial capacity to support their child's education. The respondent is uncertain about how to afford the costs of their children's education, which indicates a lack of financial resources. Label: Low Budget

Input:
Interviewer: How do you plan to support your child's education? Respondent: We're doing our best to save money, but it's tough. We might have to consider taking out a loan.

Output:
Interpretation: The respondent mentions the difficulty in saving money and the consideration of a loan, which suggests a struggle with financial constraints.
This indicates a lack of financial resources to support their child's education. Label: Low Budget

Input:
Interviewer: How do you see your child's future? Respondent: I see my child becoming a doctor. We're doing all we can to support this dream.

Output:
Interpretation: The respondent expresses a hopeful vision for their child's future and their commitment to support it. However, there is no explicit mention of financial constraints or limitations in their ability to support their child's education. Label: Not Applicable

Input:
Interviewer: What are your plans for your child's education? Respondent: We've been exploring different schools and educational programs. We want the best for our child.

Output:
Interpretation: The respondent talks about exploring different education options for their child, indicating a proactive approach. There is no explicit mention of a lack of financial resources or constraints in supporting their child's education. Label: Not Applicable

High Education
System prompt Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their educational aspirations for their children's futures. Specifically, your role is to discern whether the parent explicitly expresses an intention or desire for their child to achieve a high level of education, either secular or religious, regardless of whether the level is explicitly specified.
Assign the "High Education" label only if the parent explicitly articulates a desire or intention for their child to pursue a high level of education in the future.
If there is no such direct reference to high education in the parent's statement, assign the "Not Applicable" label to indicate that the topic of high education was not addressed.

Input:
Interviewer: What are your aspirations for your children's future? Respondent: I have always dreamt of them becoming successful in their respective fields. I want my eldest son to pursue medicine and my daughter to become an engineer.

Output:
Interpretation: The respondent clearly articulates a desire for their children to pursue careers in fields that require high education, such as medicine and engineering. Therefore, the dialogue aligns with the "High Education" label. Label: High Education

Input:
Interviewer: Do you have any specific goals for your children's education? Respondent: Yes, I want them to study hard and excel in their academics. I hope they can attend university and get a good degree.

Output:
Interpretation: The respondent explicitly states a desire for their children to attend university and achieve a degree, which is considered high education. Therefore, the dialogue aligns with the "High Education" label. Label: High Education

Input:
Interviewer: What are your plans for your children's education? Respondent: I just want them to be happy and healthy. Whatever they choose to do in life, I will support them.

Output:
Interpretation: Although the respondent expresses a supportive attitude towards their children's future plans, there is no explicit mention or intention of high education. Therefore, the dialogue aligns with the "Not Applicable" label. Label: Not Applicable

Input:
Interviewer: What are your dreams for your children's future? Respondent: I want them to be good people, to be respectful and kind. I hope they find happiness in whatever they choose to do.

Output:
Interpretation: While the respondent expresses aspirations for their children's character and happiness, there is no explicit mention of a desire for high education. Therefore, the dialogue aligns with the "Not Applicable" label. Label: Not Applicable

Low Education
System prompt Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their educational aspirations for their children's futures. Specifically, your role is to discern whether the parent explicitly expresses an intention or desire for their child to achieve a low level of education, either secular or religious, equivalent to or lower than higher secondary education. Do not assign this label if the child is currently studying below lower or secondary education, or if the scale of education is unspecified.
Assign the "Low Education" label only if the parent explicitly articulates a desire or intention for their child to pursue a level of education that is equivalent to or lower than higher secondary education in the future.
If there is no such direct reference to low education in the parent's statement, assign the "Not Applicable" label to indicate that the topic of low education was not addressed.

Input:
Interviewer: What are your aspirations for your son's education? Respondent: I want him to finish his schooling up to the 10th grade and then start working.

Output:
Interpretation: The respondent explicitly states that they want their child to complete education up to the 10th grade, which is equivalent to or lower than higher secondary education. Therefore, the dialogue should be assigned the "Low Education" label. Label: Low Education

Input:
Interviewer: Do you have any educational goals for your child? Respondent: Yes, I want my child to pursue a degree in medicine.

Output:
Interpretation: The respondent expresses a desire for their child to pursue a degree in medicine, which is a higher level of education than higher secondary education. Therefore, the dialogue should be assigned the "Not Applicable" label. Label: Not Applicable

Input:
Interviewer: What do you wish for your daughter's future? Respondent: I hope she completes her education till grade 8 and then gets married.

Output:
Interpretation: The respondent expresses a desire for their daughter to finish education up to the 8th grade, which is lower than higher secondary education. This aligns with the "Low Education" label. Label: Low Education

Input:
Interviewer: What are your hopes for your son's future? Respondent: I hope he becomes a successful businessman.

Output:
Interpretation: The respondent expresses a hope for their son's future but does not specify a level of education. Therefore, the dialogue should be assigned the "Not Applicable" label. Label: Not Applicable

Neutral Education
System prompt Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their educational aspirations for their children's futures. Specifically, your role is to discern whether the parent expresses a vague or general desire for their child's education, accompanied by a future job or religious education ambition, without implying a high or low level of education.
Assign the "Neutral Education" label only if the parent articulates a general or ambiguous desire for their child to study in the future, accompanied by the mention of a future job or religious education ambition, without specifically indicating a high or low level of education.
If there is no such direct reference to neutral education in the parent's statement, assign the "Not Applicable" label to indicate that the topic of neutral education was not addressed.

Input:
Interviewer: What are your aspirations for your daughter's future? Respondent: My wish is for her to receive a good education and perhaps find a good job in the future.

Output:
Interpretation: The respondent expresses a general desire for their child to receive education and hints at a future job ambition. However, there is no specific mention of a high or low level of education. Label: Neutral Education

Input:
Interviewer: What are your expectations for your child's education? Respondent: I want my child to be an engineer.

Output:
Interpretation: The respondent expresses a specific career ambition for their child, which implies a high level of education. This does not align with the "Neutral Education" label. Label: Not Applicable

Input:
Interviewer: What are your dreams for your son's future? Respondent: I want him to be successful in whatever he chooses to do, be it in his education or career.

Output:
Interpretation: The respondent expresses a general desire for their child's success in education and career but doesn't specify a level of education. Label: Neutral Education

Input:
Interviewer: How do you envision your child's future? Respondent: I just want them to be happy and healthy. If they choose to pursue education or a career, that's up to them.

Output:
Interpretation: The respondent expresses a general wish for their child's happiness and health, but doesn't specify any educational or career ambitions. This does not align with the "Neutral Education" label. Label: Not Applicable

Religious Education
System prompt Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their aspirations for their children's future education with a particular focus on religious education. Specifically, your role is to discern whether the parent explicitly expresses an intention or desire for their child to pursue religious education or become a religious figure in the future. Do not assign this label if the child is already engaged in any of the mentioned activities, as current religious education is not considered a future ambition.
Assign the "Religious Education" label only if the parent explicitly articulates a desire or intention for their child to engage in religious education or assume a religious role in the future.

Input:
Interviewer: What are your aspirations for your youngest son? Respondent: I want him to become a doctor, but I also want him to get a religious education. I have plans to enroll him in a Madrasa when he's a bit older.

Output:
Interpretation: The parent explicitly expresses a desire for their child to pursue religious education in the future, alongside becoming a doctor. This aligns with the "Religious Education" label. Label: Religious Education

Input:
Interviewer: How do you envision your son's future? Respondent: I want him to be a scholar of our religion.
He should learn and teach our holy scriptures.

Output:
Interpretation: The parent expresses a clear intention for their child to engage in religious education and even become a religious figure in the future. This aligns with the "Religious Education" label. Label: Religious Education

Input:
Interviewer: What are your dreams for your daughter's future? Respondent: I hope she becomes a successful engineer. I want her to have a good education and a stable career.

Output:
Interpretation: While the parent has clear aspirations for their child's future, there is no explicit mention of religious education or the desire for the child to take on a religious role. Therefore, the "Religious Education" label is not applicable. Label: Not Applicable

Input:
Interviewer: What are your plans for your children's education? Respondent: I want them to study hard and go to university. A good education is the key to a better life.

Output:
Interpretation: The parent talks about their child's education, but there is no explicit reference to religious education or the child assuming a religious role in the future. Therefore, the "Religious Education" label is not applicable. Label: Not Applicable Entrepreneur System prompt Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their aspirations concerning their children's futures. Specifically, your role is to discern whether the parent explicitly expresses any intention or wishes related to their child's potential future entrepreneurship.
Assign the "Entrepreneur" label only if the parent explicitly articulates a desire or intention for their child to pursue a non-wage enterprise job or start and run their own business, regardless of whether the business type is specified.
If there is no such direct reference to entrepreneurship in the parent's statement, assign the "Not Applicable" label to indicate that the topic of entrepreneurship was not addressed.

Input:
Interviewer: What are your aspirations for your daughter's future? Respondent: I want her to be independent and successful. I am thinking of setting up a small business for her, maybe a bakery or a boutique.

Output:
Interpretation: The respondent explicitly expresses a desire for their child to run a business in the future. This aligns with the definition of the "Entrepreneur" label. Label: Entrepreneur

Input:
Interviewer: What are your dreams for your children's future? Respondent: I want them to be happy and healthy. I hope they get a good education and find jobs that they love.

Output:
Interpretation: While the respondent talks about their children's future, there is no direct reference to entrepreneurship or starting a business. Therefore, this does not meet the criteria for the "Entrepreneur" label. Label: Not Applicable

Input:
Interviewer: What do you hope your son will do in the future? Respondent: I want him to be a successful person. I am encouraging him to start his own tech firm.

Output:
Interpretation: The respondent has directly expressed their desire for their child to start and run their own business in the future, which fits the criteria for the "Entrepreneur" label. Label: Entrepreneur

Input:
Interviewer: What do you envision for your child's future? Respondent: I want my child to be a doctor. I believe that is a noble profession and I want him to serve the community.

Output:
Interpretation: The respondent has expressed a specific desire for their child to become a doctor, which is distinct from entrepreneurship. This does not meet the criteria for the "Entrepreneur" label. Label: Not Applicable

Secular Job
System prompt Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their aspirations concerning their children's future careers in secular jobs. Specifically, your role is to discern whether the parent explicitly expresses a desire or intention related to their child pursuing a specific secular job, occupation, or work type, such as a doctor, government job, NGO job, or teacher in a non-religious school.
Assign the "Secular Job" label only if the parent explicitly articulates a desire or intention for their child to pursue a particular secular career in the future.
If there is no such direct reference to a secular job in the parent's statement, or if there is a reference to a religious job instead, assign the "Not Applicable" label to indicate that the topic of secular jobs was not addressed.

Input:
Interviewer: What are your hopes for your daughter's future? Respondent: I hope she becomes a successful lawyer. She has always been good at debating and I believe she would excel in that field.

Output:
Interpretation: The parent explicitly expresses a desire for their child to pursue a career as a lawyer, which is a secular job. Therefore, the "Secular Job" label should be assigned. Label: Secular Job

Input:
Interviewer: What are your aspirations for your son's future? Respondent: I just want him to be happy and healthy. Whatever career he chooses, I will support him.

Output:
Interpretation: While the parent expresses a general desire for their child's well-being and happiness, there is no explicit mention of a specific secular job or career. Therefore, the "Not Applicable" label should be assigned. Label: Not Applicable

Input:
Interviewer: Do you have any specific career in mind for your daughter? Respondent: Yes, I would like her to pursue a career in engineering. She has always been fascinated by how things work.

Output:
Interpretation: The parent explicitly expresses a desire for their child to pursue a career in engineering, which is a secular job. Therefore, the "Secular Job" label should be assigned. Label: Secular Job

Input:
Interviewer: What are your dreams for your child's future? Respondent: My dream is for my child to be a good person and contribute positively to society.

Output:
Interpretation: While the parent expresses a general desire for their child's future, there is no explicit mention of a specific secular job or career. Therefore, the "Not Applicable" label should be assigned. Label: Not Applicable

System prompt
Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their aspirations concerning their children's futures. Specifically, your role is to discern whether the parent explicitly expresses any intention or wishes related to their child's potential future marriage.
Assign the "Marriage" label only if the parent articulates a desire or intention for their child to get married in the future.
If there is no such reference to marriage in the parent's statement, assign the "Not Applicable" label to indicate that the topic of marriage was not addressed.

Input:
Interviewer: What are your aspirations for your son, Ahmed? Respondent: My son is very smart, I hope he will get a good job in the future. And when he is settled, I would like him to marry a good woman.

Output:
Interpretation: The respondent explicitly mentions their desire for their son to get married in the future, after getting a good job. This directly relates to the child's potential future marriage, so the label "Marriage" should be assigned. Label: Marriage

Input:
Interviewer: What do you hope for your eldest son, Omar? Respondent: Omar is a good boy. I hope he will find a good wife and start a family of his own.

Output:
Interpretation: The respondent explicitly mentions their hope for their son to find a good wife and start a family, which directly relates to the child's potential future marriage. Therefore, the label "Marriage" should be assigned. Label: Marriage

Input:
Interviewer: Can you tell me about your dreams for your daughter's future? Respondent: I want my daughter to become a successful lawyer. She's always had a passion for justice.

Output:
Interpretation: The respondent mentions their aspirations for their daughter's future, but it is related to her career, not her marital status. Since there is no explicit mention of marriage, the label "Not Applicable" should be assigned. Label: Not Applicable

Input:
Interviewer: What are your dreams for your children? Respondent: I want them to be happy and successful in their respective fields. I have always encouraged them to pursue their passions.

Output:
Interpretation: The respondent talks about their aspirations for their children's happiness and success, but there is no explicit mention of marriage. Therefore, the label "Not Applicable" should be assigned. Label: Not Applicable Migration System prompt Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their aspirations concerning their children's futures. Specifically, your role is to discern whether the parent explicitly expresses any intention or wishes related to their child's potential future migration.
Assign the "Migration" label only if the parent explicitly articulates a desire or intention for their child to migrate in the future for reasons such as work, studying, or resettling.
If there is no such direct reference to migration in the parent's statement, assign the "Not Applicable" label to indicate that the topic of migration was not addressed.

Input:
Interviewer: What are your thoughts on your daughter's future after she completes her studies? Respondent: I am hoping to send her to a foreign university for her post-graduation studies. I believe it will provide her with better opportunities.

Output:
Interpretation: The parent is explicitly expressing a desire for their child to migrate for further studies. This aligns with the "Migration" label as the parent is directly referring to the child's potential future migration. Label: Migration

Input:
Interviewer: Have you considered options for your daughter's higher education? Respondent: Yes, I am considering sending her abroad for her Master's degree. I think she will benefit from the exposure and opportunities there.

Output:
Interpretation: The parent is explicitly expressing a desire for their child to migrate for higher education. This aligns with the "Migration" label as the parent is directly referring to the child's potential future migration. Label: Migration

Input:
Interviewer: What are your aspirations for your son's future? Respondent: I want him to have a good education and a successful career.
I hope he becomes a respected doctor.

Output:
Interpretation: The parent is discussing their child's future but there is no explicit mention of migration. The parent is expressing aspirations related to the child's career but not necessarily in the context of moving abroad. Hence, this does not align with the "Migration" label. Label: Not Applicable

Input:
Interviewer: What are your plans for your son after he completes his high school? Respondent: I want him to enrol in a good university here and study engineering.
We have excellent universities in our country.

Output:
Interpretation: The parent is discussing their child's future but there is no explicit mention of migration. The parent is expressing aspirations related to the child's education but it is specifically mentioned to be within their own country. Hence, this does not align with the "Migration" label. Label: Not Applicable

No Ambition
System prompt Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their aspirations for their children's futures. Specifically, your role is to discern whether the parent expresses helplessness, unwillingness, or a lack of dreams or plans for their child's future.
Assign the "No Ambition" label only if the parent explicitly conveys a lack of ambitions or plans for their child's future or if they demonstrate a sense of helplessness or unwillingness concerning their child's future.
If there is no such direct expression of a lack of ambition in the parent's statement, assign the "Not Applicable" label to indicate that the topic of ambition was not addressed.

Input:
Interviewer: What are your hopes for your daughter's future? Respondent: What future? We are just surviving. She will just have to get married and that's it.

Output:
Interpretation: The respondent expresses a sense of helplessness regarding their daughter's future and does not express any dreams or plans for her future beyond getting married. This aligns with the "No Ambition" label.

Input:
Interviewer: What are your plans for your child's education? Respondent: I can't afford to send him to school. He'll have to work like the rest of us.

Output:
Interpretation: The respondent expresses a sense of helplessness and lack of plans for their child's education, indicating a lack of ambition for the child's future. This aligns with the "No Ambition" label.

Input:
Interviewer: What are your dreams for your son's future? Respondent: I want him to become a doctor and serve the community.

Output:
Interpretation: The respondent clearly expresses a dream for their child's future, indicating that they have ambitions for the child. This does not align with the "No Ambition" label. Label: Not Applicable

Input:
Interviewer: How do you see your child's future? Respondent: I hope he will be able to get a good education and have a successful career.

Output:
Interpretation: The respondent expresses hopes for their child's future, indicating that they have ambitions for the child. This does not align with the "No Ambition" label. Label: Not Applicable

Reliance on God
System prompt Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their reliance on divine intervention or faith in God to support their children's education and future aspirations. Specifically, your role is to discern whether the parent explicitly expresses full or partial reliance on God to fulfill their dreams for their children or if they currently rely on God's will in their lives.
Assign the "Reliance on God" label only if the parent explicitly articulates a full or partial dependence on God to accomplish their aspirations for their children's future or if they demonstrate a current reliance on divine will.
If there is no such direct reference to reliance on God in the parent's statement, assign the "Not Applicable" label to indicate that the topic of reliance on divine intervention was not addressed.

Input:
Interviewer: What are your hopes for your children's future? Respondent: I hope they will do well in their studies and become successful. But ultimately, it's all up to God's will.

Output:
Interpretation: The respondent clearly states their hopes for their children's future, but also expresses a reliance on God's will to determine the outcome. This demonstrates a partial reliance on divine intervention for their children's future. Label: Reliance on God

Input:
Interviewer: Do you have any dreams for your daughter's future? Respondent: Yes, I dream that she will become a doctor. But we leave everything in God's hands. If it's His will, it will happen.

Output:
Interpretation: The respondent expresses a dream for their daughter's future but also mentions that the fulfillment of this dream is dependent on God's will. This shows a reliance on divine intervention for their child's future. Label: Reliance on God Input: Interviewer: How do you plan to support your children's education? Respondent: I work hard every day to provide for their needs. I believe in the power of hard work and dedication.

Output: Interpretation:
The respondent talks about supporting their children's education through their own efforts and does not mention any reliance on God or divine intervention. Therefore, the "Reliance on God" label does not apply in this case. Label: Not Applicable

Input:
Interviewer: What are your plans for your son's education? Respondent: I just want him to do his best. The rest is up to him.

Output:
Interpretation: The respondent expresses a desire for their son to do well in his education, but does not mention any reliance on God or divine intervention. Therefore, the "Reliance on God" label does not apply in this case. Label: Not Applicable

Religious Aspirations
System prompt Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their aspirations concerning their children's future education and careers solely in the context of religion, without implying any monetary incentives. Specifically, your role is to discern whether the parent explicitly expresses a desire for their child to pursue a religious path.
Assign the "Religious Aspiration" label only if the parent explicitly articulates aspirations for their child that include becoming a religious scholar, Hafiz, attending a madrassa, learning Arabic, Quran reading, Islamic covering, regular prayer, working in Islamic banks, or being a great Islamic scholar. It is crucial not to assign this label if the child is already engaged in any of the mentioned activities, as current religious activities are not considered future aspirations.
If there is no such direct reference to religious aspirations in the parent's statement, or if there is a reference to secular aspirations, assign the "Not Applicable" label to indicate that the topic of religious aspirations was not addressed.

Input:
Interviewer: What are your aspirations for your son's future? Respondent: I hope that he will become a great Islamic scholar. I want him to learn Arabic and read the Quran regularly.

Output:
Interpretation: The parent explicitly expresses a desire for their child to pursue a path in religious education, specifically mentioning becoming an Islamic scholar, learning Arabic, and reading the Quran regularly. Therefore, this conversation aligns with the "Religious Aspiration" label. Label: Religious Aspiration

Input:
Interviewer: What do you envision for your child's future? Respondent: I would like my son to become a Hafiz. It's a great honor in our religion.

Output:
Interpretation: The parent explicitly expresses a desire for their child to become a Hafiz, which is a religious aspiration as per the instruction. Therefore, this conversation aligns with the "Religious Aspiration" label. Label: Religious Aspiration

Input:
Interviewer: Do you have any specific dreams for your daughter? Respondent: I want her to excel in her studies and become a successful engineer.

Output:
Interpretation: While the parent does express aspirations for their child's future, these aspirations are not religious in nature. The parent hopes for their child to become an engineer, which doesn't involve any religious activities as per the instruction. Therefore, this conversation should be labeled as "Not Applicable". Label: Not Applicable

Input:
Interviewer: What are your hopes for your children's future careers? Respondent: I want them to have successful careers. I hope my daughter becomes a doctor and my son becomes a lawyer.

Output:
Interpretation: While the parent does express aspirations for their children's future, these aspirations do not involve any religious activities or pursuits as per the instruction. Therefore, this conversation should be labeled as "Not Applicable". Label: Not Applicable

Secular Aspirations
System prompt Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their aspirations concerning their children's future education, career, and personal development without religious context. Specifically, your role is to discern whether the parent explicitly expresses a desire for their child to develop positive character traits, moralistic values, or general well-being.
Assign the "Secular Aspiration" label only if the parent explicitly articulates aspirations for their child that include becoming a good person, being responsible, having good manners, living a healthy life, achieving professional success, or having a good family.
If there is no such direct reference to secular aspirations in the parent's statement, or if there is a reference to religious aspirations, assign the "Not Applicable" label to indicate that the topic of secular aspirations was not addressed.

Input:
Interviewer: Can you share your aspirations for your children? Respondent: I want them to be kind, respectful and educated individuals. I hope they lead a healthy life and contribute positively to society.

Output:
Interpretation: The respondent explicitly expresses a desire for their children to develop positive traits such as kindness and respectfulness.
They also mention the importance of education and leading a healthy life, which are secular aspirations for their children. Therefore, the dialogue should be assigned the "Secular Aspiration" label. Label: Secular Aspiration

Input:
Interviewer: What are your dreams for your children? Respondent: My dream for my children is that they grow up to be responsible, hardworking, and successful in their careers. I also hope they have a happy family life.

Output:
Interpretation: The respondent clearly articulates secular aspirations for their children, which include becoming responsible and hardworking individuals, achieving professional success, and having a good family life. Therefore, the dialogue should be assigned the "Secular Aspiration" label. Label: Secular Aspiration

Input:
Interviewer: What kind of future do you envision for your children? Respondent: I just hope they're happy, whatever they choose to do.

Output:
Interpretation: While the respondent expresses a desire for their children's happiness, they do not explicitly state any secular aspirations such as developing positive character traits or achieving professional success. Therefore, the dialogue should be assigned the "Not Applicable" label. Label: Not Applicable

Input:
Interviewer: What kind of person would you like your child to become and how do you want them to live their life? Respondent: I hope they will follow our religious teachings, pray regularly, and become a great Islamic scholar.

Output:
Interpretation: The respondent discusses faith-based aspirations for their child, such as following religious teachings, praying regularly, and becoming a great Islamic scholar. Therefore, the dialogue should be assigned the "Not Applicable" label. Label: Not Applicable

Vocational Training
System prompt Your task involves interpreting and analyzing transcripts of interviews conducted with parents, focusing on their aspirations concerning their children's future vocational training. Specifically, your role is to discern whether the parent explicitly expresses a desire or intention related to their child pursuing vocational training or learning a specific skill in the context of ambition, such as tailoring, handicrafts, or sewing. The label should only be assigned if training, teaching, or learning is mentioned in the context of ambition.
Assign the "Vocational Training" label only if the parent explicitly articulates a desire or intention for their child to undergo vocational training or learn a particular skill in the future as part of their ambition.
If there is no such direct reference to vocational training in the parent's statement, assign the "Not Applicable" label to indicate that the topic of vocational training was not addressed.

Input:
Interviewer: What are your aspirations for your youngest daughter? Respondent: I want her to learn sewing. It's a skill that can help her in the future.

Output:
Interpretation: The parent explicitly mentions a desire for their daughter to learn sewing, a vocational skill. Therefore, the dialogue should be labeled as 'Vocational Training'. Label: Vocational Training

Input:
Interviewer: How do you envision your child's future? Respondent: Well, I believe in letting him choose his own path. But I do encourage him to learn carpentry. It's a useful skill to have.

Output:
Interpretation: The parent expresses a desire for their child to learn carpentry, a vocational skill. Therefore, the dialogue should be labeled as 'Vocational Training'. Label: Vocational Training

Input:
Interviewer: Do you have any specific plans for your son's future? Respondent: I hope he continues his studies and becomes a doctor one day.

Output:
Interpretation: Although the parent expresses a desire for their child's future, there is no mention of vocational training or learning a specific skill. Therefore, the dialogue should be labeled as 'Not Applicable'. Label: Not Applicable

Input:
Interviewer: What are your dreams for your children? Respondent: I just want them to be happy and healthy. That's all that matters.

Output:
Interpretation: The parent expresses a general wish for their children's well-being, but there is no explicit mention of vocational training or learning a specific skill. Therefore, the dialogue should be labeled as 'Not Applicable'. Label: Not Applicable

A.1 LLM prompts for Data Augmentation
We directed the LLM with the following "system prompt": Revise a given dialogue between an interviewer and respondent while maintaining the core meaning. Additionally, you may alter names, places, and other identifying information while preserving the core meaning of the dialogue.
Create a single revision of the dialogue in the following format: Interviewer: (question/statement by the interviewer) Respondent: (answer by the respondent)

B iQual Appendix
This Appendix summarises the key features of the iQual approach to supervised NLP for qualitative analysis. The method is set out in more detail in Ashwin et al. (2022) and available through a Python package at https://github.com/worldbank/iQual.

B.1 Summary of Approach
The overall goal of our iQual is to use our subset of annotated interviews to provide reliable annotations for the remainder of the sample. Broadly, we do this by training a series of classifier models on our annotated set and then using this model to predict annotations for the unannotated set. We train and predict for each annotations separately, so the model for Religious Aspiration will be trained and make its predictions separately from the model for Secular Aspiration. As the qualitative annotations are defined at the level of question-answer pairs (QA), we can represent each annotation as a binary classification problem at the QA level. Figure 6, also shown in Ashwin et al. (2022), illustrates our overall methodology for a single annotation. On the left hand side we see a "human" sample of size N h , in which interviews include both text w and annotations y, and a "machine" sample in which interviews include only the text. As annotations are defined at the QA (question-answer pair) level, so we represent w h i,s as the sth QA in interview i in the human sample, with y h i,s being the binary annotation on that QA. In other words, if the annotation Religious Aspiration, y h i,s will be equal to one if that QA has been annotated as displaying religious aspirations, and will be zero otherwise.
Aggregate to interview level Aggregate to interview level

Substantive analysis
We then train some classifier f () parameterised by θ to predict y h i,s based on the QA text w h i,s . As we will discuss below, there are many options for both the classifier we can use here, as well as how to represent the text numerically. A key point here is that the text representation must by full unsupervised -i.e. we do not use any information about y or any further information about the interview subject when creating a numerical representation of the text. The text representation, classifier and a variety of hyperparameters are chosen using k-fold cross-validation. Given this trained classifier we can then predict annotations a the QA level for our unannotated "machine" sample. This gives us the predicted annotationsŷ m i,s . We then aggregate the QA level annotations to the interview level using aggregation function g().
The choice of this aggregation function is at least in part a substantive question that depends on the research question. For example, if we take the mean value of y across QA pairs for each interview this gives us a measure of the intensity with which this concept comes up. On the other hand, if we take the maximum value across the interview this gives us a measure of interviews in which this concept comes up at least once. We perform this aggregation for both the observed human annotations Y h , the "in-sample" predicted human annotationsŶ h and the "out-of-sample" predicted machine annotationŝ Y m . The predicted annotations for the human sample can then by used to assess the measurement errors introduced by the model. For the quantification of measurement errors, we also make use of bootstrapping. The observed human annotations and machine annotations are then combined to give an enhanced sampleỸ . Once we have verified that the enhancement does indeed add value, we proceed with substantive analysis.

B.2 Choices
There are many possible options for the numerical representation of the text representation w, the classifier f () and the aggregation function g(). We fix the aggregation function as taking the mean across QA pairs and we use cross-validation to select the text representation and the classifier. As we train the classifier for each annotation independently, this allows for the fact that a different classification model of text representation may be optimal for different annotations. THe following gives an exhaustive list of the text representations, models and hyperparameters that are selected over during cross validation. In addition to these options we also cross-validate whether the question is included along with the answer in the text representation.

TfidfVectorizer
TfidfVectorizer is a method for converting text into numerical representations, specifically term frequencyinverse document frequency (TF-IDF) vectors. It counts the frequency of words in a document and downweights the importance of commonly used words. This can be useful for text classification tasks, as it allows the classifier to focus on the words that are most distinctive to a particular document.
• ngram_range: The range of n-grams to consider when creating the vocabulary.
• min_df : The minimum number of documents a word must be in to be included in the vocabulary.
• max_df : The maximum number of documents a word can be in to be included in the vocabulary.
• max_features: The maximum number of words to keep in the vocabulary, based on word frequency.
• use_idf : A boolean flag indicating whether to use the inverse-documentfrequency weighting.
• norm: The type of normalization to apply to the vectors.
• smooth_idf : A boolean flag indicating whether to smooth the idf values.
• sublinear_tf : A boolean flag indicating whether to apply sublinear scaling to the term frequency.

CountVectorizer
CountVectorizer is a method for converting text into numerical representations, specifically a sparse matrix of word counts.
It counts the frequency of words in a document and does not down-weight the importance of commonly used words. This can be useful for text classification tasks, as it allows the classifier to consider all words equally, rather than down-weighting the importance of commonly used words.
• ngram_range: The range of n-grams to consider when creating the vocabulary.
• min_df : The minimum number of documents a word must be in to be included in the vocabulary.
• max_df : The maximum number of documents a word can be in to be included in the vocabulary.  This a pre-trained language understanding model that combines the advantages of masked language modeling (MLM) and permuted language modeling (PLM) to address the limitations of both methods. It leverages the dependency among predicted tokens through PLM and takes auxiliary position information as input to make the model see a full sentence, reducing the position discrepancy between pre-training and fine-tuning. This model was pre-trained on a large-scale dataset and generates a vector of 768 dimensions.
all-roberta-large-v1 1024 This is a pre-trained language understanding model with a vector representation of 1024 dimensions. It was developed as an improvement upon the BERT model and was trained using the masked language modeling (MLM) objective. It has achieved strong performance on natural language processing tasks and can be finetuned on labeled datasets for specific tasks such as classification or language translation. average_word_embeddings _glove.6B.300d 300 This is a method for converting text into numerical representations, specifically word embeddings. It uses a pre-trained GloVe model to generate 300-dimensional vector representations for each word in a document, and then averages these vectors to create a single representation for the entire document. This can be useful for text classification tasks, as it allows the classifier to consider the semantic relationships between words, rather than just their frequencies.
distiluse-base-multilingualcased-v2 512 This is a pre-trained language understanding model that maps text into a 512-dimensional vector representation. It is a smaller and faster version of the popular transformer model, BERT, and has been trained on a large multilingual dataset, allowing it to process text in multiple languages. It has also been cased, meaning it can distinguish between upper and lower case letters. This model is useful for natural language processing tasks such as language translation and text classification, and can be fine-tuned on labeled datasets for specific tasks. LogisticRegression This is a linear classifier that uses a logistic function to predict the probability of a sample belonging to a particular class.
It is commonly used for binary classification tasks, but can also be used for multiclass classification by implementing a oneversus-rest approach. SGDClassifier This is a linear classifier that uses stochastic gradient descent to learn the parameters of the model. The modified huber loss function is a smooth approximation of the hinge loss, which is commonly used for linear classification tasks.
• loss: The loss function to use, with options such as "hinge", "log", "modified_huber", "squared_hinge", and "perceptron". • penalty: The type of regularization to use, with options such as L1, L2, "elasticnet", and "none". • alpha: The regularization strength, with higher values indicating stronger regularization. • l1_ratio: The proportion of L1 regularization to use in the elasticnet penalty. • tol: The tolerance for the stopping criteria. • learning_rate: The learning rate for the optimization algorithm, with options such as "constant", "optimal", and "invscaling". • eta0: The initial learning rate for the "constant" and "invscaling" learning rate schedules. • power_t: The exponent for the "invscaling" learning rate schedule.
• loss: The loss function to use.
• learning_rate: The learning rate schedule to use. ("optimal") • alpha: The constant that multiplies the regularization term. [0.00002, 1000] RandomForestClassifier This is an ensemble classifier that uses multiple decision trees to make predictions. It randomly selects a subset of features to consider at each split in the tree, which helps to reduce overfitting and improve the generalization of the model.
• n_estimators: The number of decision trees in the forest. • criterion: The function to measure the quality of a split, with options such as "gini" and "entropy". • max_depth: The maximum depth of the decision tree.
• min_samples_split: The minimum number of samples required to split an internal node.
• min_samples_leaf: The minimum number of samples required to be at a leaf node. • min_weight_fraction_leaf: The minimum weighted fraction of the sum total of weights required to be at a leaf node. • max_features: The number of features to consider when looking for the best split. • max_leaf_nodes: The maximum number of leaf nodes in the tree. • min_impurity_decrease: The minimum decrease in impurity required to split the node. • bootstrap: A boolean flag indicating whether to use bootstrap samples when building the trees. • oob_score: A boolean flag indicating whether to use outof-bag samples to estimate the generalization error.
• n_estimators: The number of trees in the forest. [100, 1000] • max_depth: The maximum depth of the tree. [10,100] DecisionTreeClassifier This is a classifier that uses a tree structure to make decisions based on the features of a sample. At each node in the tree, the classifier considers a single feature and splits the data based on the value of that feature. The final decision is made based on the path taken through the tree.
• criterion: The function to measure the quality of a split, with options such as "gini" and "entropy". • splitter: The strategy to use when searching for a split, with options such as "best" and "random". • max_depth: The maximum depth of the tree. • min_samples_split: The minimum number of samples required to split an internal node.
• min_samples_leaf: The minimum number of samples required to be at a leaf node. • min_weight_fraction_leaf: The minimum weighted fraction of the sum total of weights required to be at a leaf node. • max_features: The number of features to consider when looking for the best split. • max_leaf_nodes: The maximum number of leaf nodes in the tree. • min_impurity_decrease: The minimum decrease in impurity required to split the node.
• max_depth: The maximum depth of the tree. [5,100] • min_impurity_decrease: A node will be split if this split induces a decrease of the impurity greater than or equal to this value. [0.00002,10000]

MLPClassifier
This is a classifier that uses a neural network with multiple layers to make predictions. It is commonly used for classification tasks and can handle both continuous and categorical data.
The number of layers and the number of units in each layer can be adjusted to fit the complexity of the task.
• hidden_layer_sizes: The number of neurons in each hidden layer. • activation: The activation function to use, with options such as "identity", "logistic", "tanh", and "relu". • solver: The algorithm to use for optimization, with options such as "lbfgs", "sgd", and "adam". • alpha: The regularization strength, with higher values indicating stronger regularization.
• batch_size: The number of samples to use in each iteration of the optimization algorithm.
• learning_rate: The learning rate for the optimization algorithm, with options such as "constant", "invscaling", and "adaptive". • learning_rate_init: The initial learning rate for the "constant" and "invscaling" learning rate schedules. • power_t: The exponent for the "invscaling" learning rate schedule.
• max_iter: The maximum number of iterations to run the optimization algorithm. • shuffle: A boolean flag indicating whether to shuffle the training data before each epoch. • tol: The tolerance for the stopping criteria. • warm_start: A boolean flag indicating whether to reuse the solution of the previous call to fit. • momentum: The momentum for the optimization algorithm. • nesterovs_momentum: A boolean flag indicating whether to use Nesterov's momentum. • early_stopping: A boolean flag indicating whether to use early stopping to terminate the optimization early. • validation_fraction: The fraction of the training data to use as validation data for early stopping. • beta_1: The beta 1 parameter for the Adam optimization algorithm.
• hidden_layer_sizes: The ith element represents the number of neurons in the ith hidden layer.
• n_neighbors: The number of neighbors to use when making a prediction. • weights: The weight function to use when making a prediction, with options such as "uniform" and "distance". • algorithm: The algorithm to use for finding the nearest neighbors, with options such as "brute" and "kd_tree". • leaf_size: The number of points at which to switch to a brute force search for the nearest neighbors. • p: The power parameter for the Minkowski distance metric. • metric: The distance metric to use, with options such as "euclidean", "manhattan", and "minkowski". • metric_params: Additional parameters for the distance metric.
• n_neighbors: Number of neighbors to use by default for kneighbors queries. [10,10000] • weights: weight function used in prediction.

SVC
This is a classifier that uses a support vector machine (SVM) to find the optimal hyperplane to separate the different classes. It is commonly used for classification tasks and can handle both continuous and categorical data. The kernel function used to project the data into a higher dimensional space can be adjusted to fit the complexity of the task.
• C: The regularization strength, with higher values indicating stronger regularization. • kernel: The kernel to use for the decision function, with options such as "linear", "poly", "rbf", "sigmoid", and "precomputed". • degree: The degree of the polynomial kernel. • gamma: The kernel coefficient for the rbf, poly, and sigmoid kernels. • coef0: The independent term in the polynomial and sigmoid kernels. • shrinking: A boolean flag indicating whether to use the shrinking heuristic. • probability: A boolean flag indicating whether to enable probability estimates. • tol: The tolerance for the stopping criteria. • class_weight: The class weights to use for unbalanced classes. • verbose: The level of verbosity in the output. • decision_function_shape: The shape of the decision function, with options such as "ovo" and "ovr".