mm_action_prediction/tools/action_evaluation.py [195:221]:
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
    )
    parser.add_argument(
        "--single_round_evaluation",
        dest="single_round_evaluation",
        action="store_true",
        default=False,
        help="Single round evaluation for hidden split",
    )
    parser.add_argument(
        "--compute_std_err",
        dest="compute_std_err",
        action="store_true",
        default=False,
        help="Computes standard error for the metrics",
    )
    parser.add_argument(
        "--record_instance_results",
        dest="record_instance_results",
        action="store_true",
        default=False,
        help="Records per instance results and save it back",
    )
    try:
        parsed_args = vars(parser.parse_args())
    except (IOError) as msg:
        parser.error(str(msg))
    main(parsed_args)
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -



mm_action_prediction/tools/response_evaluation.py [119:145]:
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
    )
    parser.add_argument(
        "--single_round_evaluation",
        dest="single_round_evaluation",
        action="store_true",
        default=False,
        help="Single round evaluation for hidden split",
    )
    parser.add_argument(
        "--compute_std_err",
        dest="compute_std_err",
        action="store_true",
        default=False,
        help="Computes standard error for the metrics",
    )
    parser.add_argument(
        "--record_instance_results",
        dest="record_instance_results",
        action="store_true",
        default=False,
        help="Records per instance results and save it back",
    )
    try:
        parsed_args = vars(parser.parse_args())
    except (IOError) as msg:
        parser.error(str(msg))
    main(parsed_args)
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -



