You are working with a large multimodal dataset that contains images and corresponding text descriptions. The text descriptions are highly variable in length and content. Which of the following techniques is MOST effective for handling this variability when training a multimodal model?
A. Ignore text descriptions that are longer than a certain threshold.
B. Truncate all text descriptions to a fixed length.
C. Create a fixed-size vocabulary and discard any words not in the vocabulary.
D. Use dynamic padding and masking to handle variable-length sequences efficiently during batch processing.
E. Pad all text descriptions to the same maximum length using a special padding token.
正解:D
解説: (Pass4Test メンバーにのみ表示されます)
質問 2:
You are tasked with optimizing a multimodal model that combines audio and text data for speech recognition. The model currently struggles with noisy audio environments. Which data augmentation technique would be MOST effective in improving the model's robustness to noise?
A. Adding Gaussian noise to the audio data.
B. Translating the text into different languages and back.
C. Normalizing the text data to lowercase.
D. Randomly masking parts of the text input.
E. Rotating the images used for visual context.
正解:A
解説: (Pass4Test メンバーにのみ表示されます)
質問 3:
You are building a multimodal generative A1 system that creates 3D models from text descriptions. The system produces accurate shapes but struggles to generate realistic textures and surface details. What approach would BEST address this limitation?
A. Train a separate texture generation network conditioned on the generated 3D shape.
B. Increase the batch size during the 3D model generation phase.
C. Increase the number of parameters in the text encoder.
D. Reduce the resolution of the generated 3D models to simplify the texture generation process.
E. Add more layers to the shape decoder.
正解:A
解説: (Pass4Test メンバーにのみ表示されます)
質問 4:
You are fine-tuning a large pre-trained language model for a specific downstream task. During training, you observe that the model performs well on the training data but generalizes poorly to the validation dat a. Which of the following strategies could help improve the model's generalization performance?
A. Increase the weight decay (L2 regularization).
B. Increase the training data size by collecting more data.
C. Increase the learning rate.
D. Decrease the learning rate.
E. Implement early stopping based on the validation loss.
正解:A,B,D,E
解説: (Pass4Test メンバーにのみ表示されます)
質問 5:
You are working with a dataset of handwritten digits and training a Variational Autoencoder (VAE) to generate new digits. After training, you observe that the generated digits are blurry and lack sharp details. Which of the following modifications could potentially improve the quality of the generated digits in your VAE?
A. Increasing the weight of the KL divergence term in the VAE loss function.
B. Decreasing the dimensionality of the latent space.
C. Using a simpler decoder architecture.
D. Increasing the capacity of the encoder and decoder networks (e.g., adding more layers or neurons).
E. Reducing the weight of the KL divergence term in the VAE loss function.
正解:D,E
解説: (Pass4Test メンバーにのみ表示されます)
Rikako -
この問題集だけを使って試験を受けたところ、見事合格できました。試験の形式は模擬試験とほぼ同じで、試験中も模擬試験をやっているようでとてもリラックスして試験を受けることができました。ありがとうございました。オススメです。