blob: 80d92ce852ae94dbb58af273e360666088ca5360 [file] [log] [blame] [edit]
#include "aidl/android/hardware/neuralnetworks/IPreparedModelParcel.h"
#include <android/binder_parcel_utils.h>
#include <aidl/android/hardware/neuralnetworks/BnBurst.h>
#include <aidl/android/hardware/neuralnetworks/BnExecution.h>
#include <aidl/android/hardware/neuralnetworks/BnFencedExecutionCallback.h>
#include <aidl/android/hardware/neuralnetworks/BnPreparedModel.h>
#include <aidl/android/hardware/neuralnetworks/BpBurst.h>
#include <aidl/android/hardware/neuralnetworks/BpExecution.h>
#include <aidl/android/hardware/neuralnetworks/BpFencedExecutionCallback.h>
#include <aidl/android/hardware/neuralnetworks/BpPreparedModel.h>
#include <aidl/android/hardware/neuralnetworks/IBurst.h>
#include <aidl/android/hardware/neuralnetworks/IExecution.h>
#include <aidl/android/hardware/neuralnetworks/IFencedExecutionCallback.h>
#include <aidl/android/hardware/neuralnetworks/IPreparedModel.h>
namespace aidl {
namespace android {
namespace hardware {
namespace neuralnetworks {
const char* IPreparedModelParcel::descriptor = "android.hardware.neuralnetworks.IPreparedModelParcel";
binder_status_t IPreparedModelParcel::readFromParcel(const AParcel* _aidl_parcel) {
binder_status_t _aidl_ret_status = STATUS_OK;
int32_t _aidl_start_pos = AParcel_getDataPosition(_aidl_parcel);
int32_t _aidl_parcelable_size = 0;
_aidl_ret_status = AParcel_readInt32(_aidl_parcel, &_aidl_parcelable_size);
if (_aidl_ret_status != STATUS_OK) return _aidl_ret_status;
if (_aidl_parcelable_size < 4) return STATUS_BAD_VALUE;
if (_aidl_start_pos > INT32_MAX - _aidl_parcelable_size) return STATUS_BAD_VALUE;
if (AParcel_getDataPosition(_aidl_parcel) - _aidl_start_pos >= _aidl_parcelable_size) {
AParcel_setDataPosition(_aidl_parcel, _aidl_start_pos + _aidl_parcelable_size);
return _aidl_ret_status;
}
_aidl_ret_status = ::ndk::AParcel_readData(_aidl_parcel, &preparedModel);
if (_aidl_ret_status != STATUS_OK) return _aidl_ret_status;
AParcel_setDataPosition(_aidl_parcel, _aidl_start_pos + _aidl_parcelable_size);
return _aidl_ret_status;
}
binder_status_t IPreparedModelParcel::writeToParcel(AParcel* _aidl_parcel) const {
binder_status_t _aidl_ret_status;
size_t _aidl_start_pos = AParcel_getDataPosition(_aidl_parcel);
_aidl_ret_status = AParcel_writeInt32(_aidl_parcel, 0);
if (_aidl_ret_status != STATUS_OK) return _aidl_ret_status;
_aidl_ret_status = ::ndk::AParcel_writeData(_aidl_parcel, preparedModel);
if (_aidl_ret_status != STATUS_OK) return _aidl_ret_status;
size_t _aidl_end_pos = AParcel_getDataPosition(_aidl_parcel);
AParcel_setDataPosition(_aidl_parcel, _aidl_start_pos);
AParcel_writeInt32(_aidl_parcel, _aidl_end_pos - _aidl_start_pos);
AParcel_setDataPosition(_aidl_parcel, _aidl_end_pos);
return _aidl_ret_status;
}
} // namespace neuralnetworks
} // namespace hardware
} // namespace android
} // namespace aidl