Risk and AI (RAI) 온라인 연습
최종 업데이트 시간: 2025년10월03일
당신은 온라인 연습 문제를 통해 GARP RAI 시험지식에 대해 자신이 어떻게 알고 있는지 파악한 후 시험 참가 신청 여부를 결정할 수 있다.
시험을 100% 합격하고 시험 준비 시간을 35% 절약하기를 바라며 RAI 덤프 (최신 실제 시험 문제)를 사용 선택하여 현재 최신 300개의 시험 문제와 답을 포함하십시오.
정답:
Explanation:
B is correct. The attention mechanism allows transformers to assign contextualized embeddings, differentiating words based on their context within the sentence
A is incorrect as the sequence-to- sequence structure refers to the model’s encoder-decoder format
C is incorrect as static embeddings do not adjust for context
D is incorrect as transformers lack recurrent connections, which are instead present in RNNs.
정답:
Explanation:
C is correct. LSTM networks are specifically designed to handle long-range dependencies by using gates to control information flow
A is incorrect as simply increasing epochs does not resolve long-range dependency issues
B is incorrect as CNNs are unsuitable for sequential text data
D is incorrect as reducing sequence length can decrease context understanding.
정답:
Explanation:
A is correct. Multimodal models are designed to generate multiple types of content, such as both text and images, based on a single prompt.
B and C are incorrect because they are limited to single content types
D is incorrect as transformers designed only for NLP cannot generate images.
정답:
Explanation:
A is correct. Naïve Bayes calculates the posterior probability for each class using word occurrences and assigns the document to the class with the highest probability.
B is irrelevant to classification
C is incorrect as probabilities differ based on words.
D ignores the role of individual word probabilities.
정답:
Explanation:
C is correct. Naïve Bayes assumes independence among features (words), meaning each word is treated independently
A is incorrect as Naïve Bayes assumes independence
B is incorrect; importance varies based on probabilities
D is incorrect because words are essential for determining class labels.
정답:
Explanation:
D is correct. Using n-grams, especially bi-grams, helps capture phrases with negation like “not helpful.”
A is incorrect, as removing stop words could discard important negation terms.
B is unrelated to capturing multi-word phrases.
C is about reducing words to their root, not handling negations.
정답:
Explanation:
B is correct. L2 normalization controls the impact of frequent words, balancing each word’s influence
A is incorrect as it would remove valuable context.
C does not reduce frequency impact but consolidates similar words.
D is irrelevant to managing word repetition.
정답:
Explanation:
B is correct. L2 normalization scales the vector, reducing the impact of any single term, even if repeated
A is incorrect as manual removal is inefficient for large datasets.
C does not address the dominance of frequent words.
D would only indicate presence or absence, not reduce the weight of repeated terms.
정답:
Explanation:
B is correct. Lemmatization reduces words to their base form, making “increase” and “increasing” identical in analysis
A is incorrect as stop word removal removes non-informative words without transforming base forms
C is incorrect because tokenization splits text into components but does not simplify word forms
D is incorrect as feature extraction converts text into numerical formats post- processing.
정답:
Explanation:
C is correct. Random search is generally faster as it samples fewer points and avoids exhaustive grid testing
A is incorrect; random search does not guarantee the exact optimal value.
B is incorrect, as random search may not outperform grid search.
D is incorrect, as random search does not test all possible values.
정답:
Explanation:
B is correct. Ridge regression (L2 regularization) reduces coefficient magnitudes without setting them to zero, making it ideal when all features should be retained
A is incorrect as LASSO can set coefficients to zero, potentially eliminating features.
C is partially correct as Elastic Net may also zero out some coefficients if L1 is significant
D is incorrect as stepwise selection does not perform regularization.
정답:
Explanation:
A is correct because opting for a simpler, interpretable model may sacrifice prediction accuracy compared to complex models
B is incorrect as interpretability is not sacrificed in this case; it’s actually prioritized
C is incorrect since simplicity is maintained, not sacrificed
D is incorrect as bias may still exist, but it is not directly due to interpretability.
정답:
Explanation:
C is correct because poor performance on new data with excellent performance on the training set indicates overfitting, where the model has memorized the training data rather than generalizing
A is incorrect as the model likely has too many parameters, not too few
B is incorrect as underfitting typically results in poor performance on both training and validation sets
D is incorrect because perfect generalization would imply good performance on both training and new data.
정답:
Explanation:
B is correct because the chain rule is used to compute the partial derivatives of the loss function with respect to each weight, enabling accurate updates
A is incorrect as initial weights are typically randomized, not derived from the chain rule
C is incorrect as the chain rule itself does not prevent overfitting
D is incorrect because the chain rule does not validate predictions but helps calculate gradients for weight updates.
정답:
Explanation:
A is correct. A high learning rate can cause oscillations in SGD, leading to unstable convergence
B is incorrect as overfitting is unrelated to convergence issues
C is incorrect because SGD can function well with smaller data batches.
D is partially correct but does not explain the root cause of the fluctuating convergence.