/************************************************************************* > File Name: hos_client_api.cpp > Author: pxz > Created Time: Thu 10 Sep 2020 03:00:23 PM CST ************************************************************************/ extern "C" { #include #include #include #include } #include #include #include #include #include #include #include #include #ifdef HOS_MOCK #include "mock/hos_mock.h" #endif #include "hos_client.h" #include "MESA_prof_load.h" #include "hos_common.h" #ifdef HOS_MESA_LOG #include "MESA_handle_logger.h" #else #define RLOG_LV_DEBUG 10 #define RLOG_LV_INFO 20 #define RLOG_LV_FATAL 30 #define MESA_create_runtime_log_handle(path, lv) \ (void *)fopen((path), "a+") #define MESA_destroy_runtime_log_handle(handle) \ fclose((FILE *)handle) #define MESA_HANDLE_RUNTIME_LOG(handle, lv, mod, fmt, args...) \ do{ \ fprintf(((FILE *) handle), "line:%d, level:%d, module:%s, ", __LINE__, lv, mod); \ fprintf(((FILE *) handle), fmt, ##args);\ fprintf(((FILE *) handle), "\n");\ } while (0) #endif #define FILE_MODE 0 #define BUFF_MODE 1 #define APPEND_MODE 2 struct hos_info_s g_hos_instance; hos_client_handle_t g_hos_handle;//一个进程只允许有一个g_hos_handle static std::mutex m_client_lock; static std::mutex m_delete_lock; static Aws::SDKOptions g_options; static inline size_t get_current_ms() { struct timespec timenow; clock_gettime(CLOCK_MONOTONIC, &timenow); return (timenow.tv_sec * 1000 + timenow.tv_nsec / 1000 / 1000 ); } static void hos_destroy_handle(hos_handle handle) { std::lock_guard locker(m_delete_lock); if (handle == NULL) { return; } put_finished_callback callback = (put_finished_callback)handle->callback; if (callback) { callback(handle->reslut, handle->bucket, handle->object, handle->error, handle->userdata); } if (handle->bucket) { free(handle->bucket); handle->bucket = NULL; } if (handle->object) { free(handle->object); handle->object = NULL; } free(handle); return ; } static void PutObjectAsyncFinished(const Aws::S3::S3Client* S3Client, const Aws::S3::Model::PutObjectRequest& request, const Aws::S3::Model::PutObjectOutcome& outcome, const std::shared_ptr& context) { const char *error = NULL; hos_func_thread_t *hos_func = &g_hos_handle.hos_func; data_info_t *data_info = NULL; const Aws::String& uuid = context->GetUUID(); size_t handle_addr = 0, stream_len = 0; hos_handle_t *handle = NULL; sscanf(uuid.c_str(), "%lu %lu", &handle_addr, &stream_len); handle = (hos_handle_t *)handle_addr; bool result = outcome.IsSuccess(); if (!result) { error = outcome.GetError().GetMessage().c_str(); MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_FATAL, __FUNCTION__, "error: [%s:%s] upload failed. error:%s", handle->bucket, handle->object, error); if (hos_func->fs2_info.fs2_handle) { data_info = (data_info_t *)&hos_func->fs2_info.data_info; atomic_add(&data_info->tx_failed_pkts, 1); atomic_add(&data_info->tx_failed_bytes, stream_len); } } else { if (hos_func->fs2_info.fs2_handle) { data_info = (data_info_t *)&hos_func->fs2_info.data_info; atomic_add(&data_info->tx_pkts, 1); atomic_add(&data_info->tx_bytes, stream_len); MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_DEBUG, __FUNCTION__, "debug: [%s:%s] upload success. tx_pkts:%lu, tx_bytes:%lu", handle->bucket, handle->object, atomic_read(&data_info->tx_pkts), atomic_read(&data_info->tx_bytes)); } else { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_DEBUG, __FUNCTION__, "debug: [%s:%s] upload success. stream size:%lu", handle->bucket, handle->object, stream_len); } } handle->reslut = result; handle->error = error; if (handle->mode & APPEND_MODE) { //APPEND MODE 保留fd atomic_add(&(handle->recive_cnt), 1); if (handle->handle_status == HOS_HANDLE_CANCEL) { if (handle->position == handle->recive_cnt) { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_DEBUG, __FUNCTION__, "debug: [%s:%s] upload completed. delete", handle->bucket, handle->object); hos_destroy_handle(handle); } } } else { //完整上传 删除fd MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_DEBUG, __FUNCTION__, "debug: [%s:%s] upload completed. handle destory", handle->bucket, handle->object); hos_destroy_handle(handle); } atomic_sub(&g_hos_handle.task_num, 1); atomic_sub(&g_hos_handle.task_context, stream_len); } static bool hos_client_create() { hos_config_t *hos_conf = &g_hos_handle.hos_config; void *log = g_hos_handle.log; if (g_hos_handle.S3Client != NULL) { g_hos_handle.count++; return true; } Aws::InitAPI(g_options); Aws::Client::ClientConfiguration config; Aws::Auth::AWSCredentials credentials(hos_conf->accesskeyid, hos_conf->secretkey); //初始化 char endpoint[128]; snprintf(endpoint, 128, "http://%s:%d/hos/", hos_conf->ip, hos_conf->port); config.endpointOverride = endpoint; config.verifySSL = false; config.enableEndpointDiscovery = true; if (hos_conf->pool_thread_size > 0) { //异步模式 //config.executor = std::shared_ptr(std::make_shared(hos_conf->pool_thread_size, Aws::Utils::Threading::OverflowPolicy::REJECT_IMMEDIATELY)); //支持线程池 config.executor = std::shared_ptr(std::make_shared(hos_conf->pool_thread_size, Aws::Utils::Threading::OverflowPolicy::QUEUE_TASKS_EVENLY_ACCROSS_THREADS)); //支持线程池 } else { //同步模式 } #ifndef HOS_MOCK g_hos_handle.S3Client = new Aws::S3::S3Client(credentials, config, Aws::Client::AWSAuthV4Signer::PayloadSigningPolicy::Never, false); #else g_hos_handle.S3Client = new Aws::S3::S3ClientMock(credentials, config, Aws::Client::AWSAuthV4Signer::PayloadSigningPolicy::Never, false); #endif /* 获取当前用户的所有的buckets */ Aws::S3::Model::ListBucketsOutcome outcome = g_hos_handle.S3Client->ListBuckets(); if (!outcome.IsSuccess()) { delete g_hos_handle.S3Client; g_hos_handle.S3Client = NULL; Aws::ShutdownAPI(g_options); g_hos_instance.error_code = (size_t)outcome.GetError().GetErrorType() + 1; snprintf(g_hos_instance.error_message, HOS_ERROR_MESSAGE_SIZE, outcome.GetError().GetMessage().c_str()); MESA_HANDLE_RUNTIME_LOG(log, RLOG_LV_FATAL, "hos_client_create", "error: %s", g_hos_instance.error_message); return false; } g_hos_handle.buckets = outcome.GetResult().GetBuckets(); g_hos_handle.count++; //g_hos_handle.executor = std::dynamic_pointer_cast(config.executor); g_hos_handle.task_num = 0; g_hos_handle.task_context = 0; MESA_HANDLE_RUNTIME_LOG(log, RLOG_LV_DEBUG, "hos_client_create", "debug: hos s3client create success, url:%s.",endpoint); return true; } bool hos_verify_bucket(const char *bucket) { if (bucket == NULL) { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_DEBUG, __FUNCTION__, "debug: bucket is null"); return false; } if (g_hos_handle.S3Client == NULL) { return false; } Aws::S3::Model::ListBucketsOutcome outcome = g_hos_handle.S3Client->ListBuckets(); if (outcome.IsSuccess()) { g_hos_handle.buckets = outcome.GetResult().GetBuckets(); for (Aws::S3::Model::Bucket& new_bucket : g_hos_handle.buckets) { if (strcmp(new_bucket.GetName().c_str(), bucket) == 0) { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_DEBUG, "hos_verify_bucket","debug: bucket:%s exits", bucket); return true; } } } else { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_FATAL, "hos_verify_bucket","error:%s", outcome.GetError().GetMessage().c_str()); } return false; } static void *fs2_statistics(void *ptr) { size_t rx_pkts_sum = 0; size_t rx_bytes_sum = 0; size_t tx_pkts_sum = 0; size_t tx_bytes_sum = 0; size_t tx_failed_bytes_sum = 0; size_t tx_failed_pkts_sum = 0; size_t cache_sum = 0; size_t req_overflow_sum = 0; size_t rx_pkts_interval = 0; size_t rx_bytes_interval = 0; size_t tx_pkts_interval = 0; size_t tx_bytes_interval = 0; size_t tx_failed_bytes_interval = 0; size_t tx_failed_pkts_interval = 0; size_t cache_interval = 0; size_t req_overflow_interval = 0; size_t rx_pkts_last = 0; size_t rx_bytes_last = 0; size_t tx_pkts_last = 0; size_t tx_bytes_last = 0; size_t tx_failed_bytes_last = 0; size_t tx_failed_pkts_last = 0; size_t cache_last = 0; size_t req_overflow_last = 0; fs2_info_t *fs2_info = NULL; hos_func_thread_t *hos_func = &g_hos_handle.hos_func; while(1) { if (hos_func->fs2_status == HOS_FS2_STOP) { break; } //pkts and bytes info rx_pkts_sum = 0; rx_bytes_sum = 0; tx_pkts_sum = 0; tx_bytes_sum = 0; tx_failed_bytes_sum = 0; tx_failed_pkts_sum = 0; cache_sum = 0; fs2_info = &hos_func->fs2_info; data_info_t *data_info = (data_info_t *)&fs2_info->data_info; rx_pkts_sum += atomic_read(&data_info->rx_pkts); rx_bytes_sum += atomic_read(&data_info->rx_bytes); tx_pkts_sum += atomic_read(&data_info->tx_pkts); tx_bytes_sum += atomic_read(&data_info->tx_bytes); tx_failed_bytes_sum += atomic_read(&data_info->tx_failed_bytes); tx_failed_pkts_sum += atomic_read(&data_info->tx_failed_pkts); cache_sum += atomic_read(&data_info->cache); rx_pkts_interval = rx_pkts_sum - rx_pkts_last; rx_bytes_interval = rx_bytes_sum - rx_bytes_last; tx_pkts_interval = tx_pkts_sum - tx_pkts_last; tx_bytes_interval = tx_bytes_sum - tx_bytes_last; tx_failed_pkts_interval = tx_failed_pkts_sum - tx_failed_pkts_last; tx_failed_bytes_interval = tx_failed_bytes_sum - tx_failed_bytes_last; cache_interval = cache_sum - cache_last; req_overflow_interval = req_overflow_sum - req_overflow_last; rx_pkts_last = rx_pkts_sum; rx_bytes_last = rx_bytes_sum; tx_pkts_last = tx_pkts_sum; tx_bytes_last = tx_bytes_sum; tx_failed_bytes_last = tx_failed_bytes_sum; tx_failed_pkts_last = tx_failed_pkts_sum; cache_last = cache_sum; req_overflow_last = req_overflow_sum; FS_operate(fs2_info->fs2_handle, fs2_info->line_ids[0], fs2_info->column_ids[0], FS_OP_SET, rx_pkts_interval); FS_operate(fs2_info->fs2_handle, fs2_info->line_ids[0], fs2_info->column_ids[1], FS_OP_SET, rx_bytes_interval); FS_operate(fs2_info->fs2_handle, fs2_info->line_ids[0], fs2_info->column_ids[2], FS_OP_SET, tx_pkts_interval); FS_operate(fs2_info->fs2_handle, fs2_info->line_ids[0], fs2_info->column_ids[3], FS_OP_SET, tx_bytes_interval); FS_operate(fs2_info->fs2_handle, fs2_info->line_ids[0], fs2_info->column_ids[4], FS_OP_SET, tx_failed_pkts_interval); FS_operate(fs2_info->fs2_handle, fs2_info->line_ids[0], fs2_info->column_ids[5], FS_OP_SET, tx_failed_bytes_interval); FS_operate(fs2_info->fs2_handle, fs2_info->line_ids[0], fs2_info->column_ids[6], FS_OP_SET, cache_interval); FS_operate(fs2_info->fs2_handle, fs2_info->line_ids[0], fs2_info->column_ids[7], FS_OP_SET, req_overflow_interval); FS_operate(fs2_info->fs2_handle, fs2_info->line_ids[1], fs2_info->column_ids[0], FS_OP_SET, rx_pkts_sum); FS_operate(fs2_info->fs2_handle, fs2_info->line_ids[1], fs2_info->column_ids[1], FS_OP_SET, rx_bytes_sum); FS_operate(fs2_info->fs2_handle, fs2_info->line_ids[1], fs2_info->column_ids[2], FS_OP_SET, tx_pkts_sum); FS_operate(fs2_info->fs2_handle, fs2_info->line_ids[1], fs2_info->column_ids[3], FS_OP_SET, tx_bytes_sum); FS_operate(fs2_info->fs2_handle, fs2_info->line_ids[1], fs2_info->column_ids[4], FS_OP_SET, tx_failed_pkts_sum); FS_operate(fs2_info->fs2_handle, fs2_info->line_ids[1], fs2_info->column_ids[5], FS_OP_SET, tx_failed_bytes_sum); FS_operate(fs2_info->fs2_handle, fs2_info->line_ids[1], fs2_info->column_ids[6], FS_OP_SET, cache_sum); FS_operate(fs2_info->fs2_handle, fs2_info->line_ids[1], fs2_info->column_ids[7], FS_OP_SET, req_overflow_sum); sleep(1); } pthread_exit(NULL); } static screen_stat_handle_t hos_init_fs2(char *app_name, int app_name_size) { int value = 0; screen_stat_handle_t fs2_handle = FS_create_handle(); hos_config_t *hos_conf = &g_hos_handle.hos_config; FS_set_para(fs2_handle, APP_NAME, app_name, app_name_size + 1); //value = 0; //true //FS_set_para(fs2_handle, FLUSH_BY_DATE, &value, sizeof(value)); if (hos_conf->fs2_path != NULL) { FS_set_para(fs2_handle, OUTPUT_DEVICE, hos_conf->fs2_path, strlen(hos_conf->fs2_path) + 1); } value = 1; FS_set_para(fs2_handle, PRINT_MODE, &value, sizeof(value)); FS_set_para(fs2_handle, CREATE_THREAD, &value, sizeof(value)); FS_set_para(fs2_handle, METRIS_FORMAT, &hos_conf->fs2_fmt, sizeof(hos_conf->fs2_fmt)); FS_set_para(fs2_handle, STAT_CYCLE, &value, sizeof(value)); value = 4096; FS_set_para(fs2_handle, MAX_STAT_FIELD_NUM, &value, sizeof(value)); if (hos_conf->fs2_ip == NULL) { FS_set_para(fs2_handle, STATS_SERVER_IP, "127.0.0.1", strlen("127.0.0.1")); } else { FS_set_para(fs2_handle, STATS_SERVER_IP, hos_conf->fs2_ip, strlen(hos_conf->fs2_ip)); } FS_set_para(fs2_handle, STATS_SERVER_PORT, &hos_conf->fs2_port, sizeof(hos_conf->fs2_port)); value = FS_OUTPUT_STATSD; FS_set_para(fs2_handle, STATS_FORMAT, &value, sizeof(value)); return fs2_handle; } static void hos_expand_fs2() { fs2_info_t *fs2_info = NULL; screen_stat_handle_t fs2_handle = NULL; hos_func_thread_t *hos_func = &g_hos_handle.hos_func; size_t i = 0; if (hos_func->fs2_info.fs2_handle) return; //data info /********************************************************************************************************** * rx_pkts rx_bytes tx_pkts tx_bytes tx_failed_p tx_failed_b cache_bytes req_overflow * current 10 100 1 100 0 0 100 0 * total 100 1000 10 1000 0 0 100(无实意) 0 ***********************************************************************************************************/ fs2_info = &hos_func->fs2_info; hos_func->fs2_info.fs2_handle = hos_init_fs2((char *)"hos-data", strlen("hos-data")); fs2_handle = hos_func->fs2_info.fs2_handle; fs2_info->line_ids = (int *)calloc(2, sizeof(int)); fs2_info->column_ids = (int *)calloc(8, sizeof(int)); const char *data_col[] = {"rx_pkts", "rx_bytes", "tx_pkts", "tx_bytes", "tx_failed_p", "tx_failed_b", "cache_bytes", "req_overflow"}; for (i = 0; i < sizeof(data_col) / sizeof(const char *); i++) { fs2_info->column_ids[i] = FS_register(fs2_handle, FS_STYLE_COLUMN, FS_CALC_CURRENT, data_col[i]); } fs2_info->line_ids[0] = FS_register(fs2_handle, FS_STYLE_LINE, FS_CALC_CURRENT, "current"); fs2_info->line_ids[1] = FS_register(fs2_handle, FS_STYLE_LINE, FS_CALC_CURRENT, "total"); hos_func->fs2_status = HOS_FS2_START; FS_start(fs2_handle); pthread_create(&hos_func->fs2_thread, NULL, fs2_statistics, NULL); return ; } static int hos_putobject_async(hos_handle_t *handle, Aws::S3::Model::PutObjectRequest& request, size_t stream_len) { char buf[128]; int ret = 0; hos_func_thread_t *hos_func = &g_hos_handle.hos_func; data_info_t *data_info = NULL; hos_config_t *hos_conf = &g_hos_handle.hos_config; //设置回调函数 std::shared_ptr context = Aws::MakeShared(""); sprintf(buf, "%lu %lu", (size_t)handle, stream_len); context->SetUUID(buf); if (hos_conf->max_request_num && hos_conf->max_request_context && (atomic_read(&g_hos_handle.task_num) >= hos_conf->max_request_num || atomic_read(&g_hos_handle.task_context) >= hos_conf->max_request_context)) { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_DEBUG, __FUNCTION__, "debug: PutObjectAsync failed. [%s:%s]. task_num:%lu, task_context:%lu", handle->bucket, handle->object, g_hos_handle.task_num, g_hos_handle.task_context); if (hos_func->fs2_info.fs2_handle) { data_info = (data_info_t *)&(hos_func->fs2_info.data_info); atomic_add(&data_info->tx_failed_pkts, 1); atomic_add(&data_info->tx_failed_bytes, stream_len); atomic_add(&data_info->tx_req_num_overflow, 1); } return HOS_SEND_FAILED; } auto &S3Client = *(g_hos_handle.S3Client); ret = S3Client.PutObjectAsync(request, PutObjectAsyncFinished, context); if (ret) { atomic_add(&g_hos_handle.task_num, 1); atomic_add(&g_hos_handle.task_context, stream_len); //不算真正成功,需要等到PutObjectAsyncFinished的结果 MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_DEBUG, __FUNCTION__, "debug: PutObjectAsync success. [%s:%s]", handle->bucket, handle->object); return HOS_CLIENT_OK; } else { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_DEBUG, __FUNCTION__, "debug: PutObjectAsync failed. [%s:%s]", handle->bucket, handle->object); if (hos_func->fs2_info.fs2_handle) { data_info = (data_info_t *)&hos_func->fs2_info.data_info; atomic_add(&data_info->tx_failed_pkts, 1); atomic_add(&data_info->tx_failed_bytes, stream_len); } return HOS_SEND_FAILED; } } static int hos_putobject_sync(hos_handle_t *handle, Aws::S3::Model::PutObjectRequest& request, size_t stream_len) { hos_func_thread_t *hos_func = &g_hos_handle.hos_func; data_info_t *data_info = NULL; auto& S3Client = *(g_hos_handle.S3Client); Aws::S3::Model::PutObjectOutcome Outcome = S3Client.PutObject(request); if (Outcome.IsSuccess()) { if (hos_func->fs2_info.fs2_handle) { data_info = (data_info_t *)&hos_func->fs2_info.data_info; atomic_add(&data_info->tx_pkts, 1); atomic_add(&data_info->tx_bytes, stream_len); MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_DEBUG, __FUNCTION__, "debug: PutObject success. [%s:%s] tx_pkts:%lu, tx_bytes:%lu", handle->bucket, handle->object, atomic_read(&data_info->tx_pkts), atomic_read(&data_info->tx_bytes)); } else { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_DEBUG, __FUNCTION__, "debug: PutObject success. [%s:%s]", handle->bucket, handle->object); } return HOS_CLIENT_OK; } else { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_DEBUG, __FUNCTION__, "debug: PutObject failed. [%s:%s] cause:%s", handle->bucket, handle->object, Outcome.GetError().GetMessage().c_str()); if (hos_func->fs2_info.fs2_handle) { data_info = (data_info_t *)&hos_func->fs2_info.data_info; atomic_add(&data_info->tx_failed_pkts, 1); atomic_add(&data_info->tx_failed_bytes, stream_len); } return (int)Outcome.GetError().GetErrorType() + 1; } } bool hos_verify_instance() { if (g_hos_handle.S3Client != NULL) { g_hos_handle.count++; return true; } memset(&g_hos_instance, 0, sizeof(g_hos_instance)); return false; } int hos_get_init_instance_errcode() { return g_hos_instance.error_code; } const char *hos_get_init_instance_errmsg() { return g_hos_instance.error_message; } const char *hos_get_url_prefix() { return g_hos_instance.hos_url_prefix; } bool hos_init_instance(const char *conf_path, const char *module, const char *bucket) { std::lock_guard locker(m_client_lock); hos_config_t *hos_conf = &g_hos_handle.hos_config; char hos_url[1024]; if (conf_path == NULL || module == NULL || bucket == NULL) { g_hos_instance.error_code = HOS_PARAMETER_ERROR; snprintf(g_hos_instance.error_message, HOS_ERROR_MESSAGE_SIZE, "param error:conf_path:%s, module:%s, bucket:%s", conf_path, module, bucket); return false; } MESA_load_profile_string_nodef(conf_path, module, "hos_serverip", hos_conf->ip, MAX_HOS_STRING_LEN); MESA_load_profile_uint_nodef(conf_path, module, "hos_serverport", &hos_conf->port); MESA_load_profile_string_nodef(conf_path, module, "hos_accesskeyid", hos_conf->accesskeyid, MAX_HOS_STRING_LEN); MESA_load_profile_string_nodef(conf_path, module, "hos_secretkey", hos_conf->secretkey, MAX_HOS_STRING_LEN); MESA_load_profile_string_def(conf_path, module, "hos_log_path", hos_conf->log_path, MAX_HOS_STRING_LEN, HOS_LOG_PATH); MESA_load_profile_uint_def(conf_path, module, "hos_log_level", &hos_conf->log_level, 30); MESA_load_profile_uint_def(conf_path, module, "hos_poolsize", &hos_conf->pool_thread_size, 0); MESA_load_profile_uint_def(conf_path, module, "hos_cache_size", &hos_conf->cache_size, 102400); MESA_load_profile_uint_def(conf_path, module, "hos_cache_count", &hos_conf->cache_count, 10); MESA_load_profile_string_def(conf_path, module, "hos_fs2_serverip", hos_conf->fs2_ip, INET6_ADDRSTRLEN, "127.0.0.1"); MESA_load_profile_uint_def(conf_path, module, "hos_fs2_serverport", &hos_conf->fs2_port, 10086); MESA_load_profile_string_def(conf_path, module, "hos_fs2_path", hos_conf->fs2_path, sizeof(hos_conf->fs2_path), "./hos_fs2.stat"); MESA_load_profile_uint_def(conf_path, module, "hos_fs2_format", &hos_conf->fs2_fmt, 0); MESA_load_profile_uint_def(conf_path, module, "hos_request_num", &hos_conf->max_request_num, 100); MESA_load_profile_uint_def(conf_path, module, "hos_request_context", &hos_conf->max_request_context, 10240000); if (strlen(hos_conf->ip) && hos_conf->port && strlen(hos_conf->accesskeyid) && strlen(hos_conf->secretkey)) { g_hos_handle.log = MESA_create_runtime_log_handle(hos_conf->log_path, hos_conf->log_level); if (g_hos_handle.log == NULL) { g_hos_instance.error_code = HOS_RUNTIME_LOG_FAILED; snprintf(g_hos_instance.error_message, HOS_ERROR_MESSAGE_SIZE, "runtime log create failed."); return false; } snprintf(hos_url, sizeof(hos_url), "http://%s:%d/hos/", hos_conf->ip, hos_conf->port); if (hos_client_create()) { //XXX 是否需要验证bucket if(hos_verify_bucket(bucket) == false) { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_FATAL, __FUNCTION__, "error: bucket:%s not exist.", bucket); hos_shutdown_instance(); g_hos_instance.error_code = HOS_BUCKET_NOT_EXIST; snprintf(g_hos_instance.error_message, HOS_ERROR_MESSAGE_SIZE, "bucket:%s not exits.", bucket); return false; } MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_DEBUG, __FUNCTION__, "debug:%s","Instance init completed"); hos_expand_fs2(); g_hos_instance.error_code = 0; g_hos_instance.error_message[0]='\0'; g_hos_instance.hos_url_prefix = (const char *)calloc(1, strlen(hos_url) + 1); memcpy((void *)g_hos_instance.hos_url_prefix, hos_url, strlen(hos_url)); return true; } else { return false; } } else { g_hos_instance.error_code = HOS_CONF_ERROR; snprintf(g_hos_instance.error_message, HOS_ERROR_MESSAGE_SIZE, "hos param error:hos ip:%s, hos port:%u, accesskeyid:%s, secretkey:%s", hos_conf->ip, hos_conf->port, hos_conf->accesskeyid, hos_conf->secretkey); return false; } } int hos_create_bucket(const char *bucket) { if (g_hos_handle.S3Client == NULL) { return HOS_INSTANCE_NOT_INIT; } if (bucket == NULL) { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_FATAL, "hos_create_bucket", "error:bucket:%s", bucket); return HOS_PARAMETER_ERROR; } auto& S3Client = *g_hos_handle.S3Client; /* 本地检查是否已经存在该bucket */ for (Aws::S3::Model::Bucket& new_bucket : g_hos_handle.buckets) { if (strcmp(new_bucket.GetName().c_str(), bucket) == 0) { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_DEBUG, __FUNCTION__, "debug: %s was exits", bucket); return HOS_CLIENT_OK; } } Aws::S3::Model::CreateBucketRequest createBucketRequest; createBucketRequest.SetBucket(bucket); Aws::S3::Model::CreateBucketOutcome createBucketOutcome = S3Client.CreateBucket(createBucketRequest); if (!createBucketOutcome.IsSuccess()) { Aws::S3::S3Errors errorcode = createBucketOutcome.GetError().GetErrorType(); if (errorcode != Aws::S3::S3Errors::BUCKET_ALREADY_OWNED_BY_YOU) { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_FATAL, __FUNCTION__,"error: %s create failed. %s", bucket, createBucketOutcome.GetError().GetMessage().c_str()); return (int)errorcode + 1; } } MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_FATAL, __FUNCTION__, "error: %s create successful", bucket); return HOS_CLIENT_OK; } static int hos_upload_stream(const char *bucket, const char *object, const char *data, size_t data_len, put_finished_callback callback, void *userdata) { data_info_t *data_info = NULL; hos_config_t *hos_conf = &g_hos_handle.hos_config; hos_func_thread_t *hos_func = &g_hos_handle.hos_func; int ret; int mode = 0; if (g_hos_handle.S3Client == NULL) { return HOS_INSTANCE_NOT_INIT; } if ((bucket == NULL) || (object == NULL)) { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_FATAL, "hos_upload_stream", "error: s3client:%s, bucket:%s, object:%s", g_hos_handle.S3Client?"not null":"null", bucket, object); return HOS_PARAMETER_ERROR; } mode = data?1:0; // 0, file mode; 1 buf mode // Create and configure the asynchronous put object request. Aws::S3::Model::PutObjectRequest request; request.SetBucket(bucket); request.SetKey(object); //设置上传数据类型 if (mode == 0) { //文件类型 const std::shared_ptr input_data = Aws::MakeShared("hos_upload_file_tag", object, std::ios_base::in | std::ios_base::binary); request.SetBody(input_data); } else { //内存块 const std::shared_ptr input_data = Aws::MakeShared("hos_upload_buf_tag"); Aws::String stream (data, data_len); *input_data << stream; request.SetBody(input_data); } //field_stat2 record if (hos_func->fs2_info.fs2_handle) { data_info = (data_info_t *)&hos_func->fs2_info.data_info; atomic_add(&data_info->rx_pkts, 1); atomic_add(&data_info->rx_bytes, data_len); } //设置回调函数 hos_handle_t *handle = (hos_handle_t *)calloc(1, sizeof(hos_handle_t)); handle->bucket = (char *)calloc(1, strlen(bucket)+1); memcpy(handle->bucket, bucket, strlen(bucket)); handle->object = (char *)calloc(1, strlen(object)+1); memcpy(handle->object, object, strlen(object)); handle->callback = callback; handle->userdata = userdata; if (hos_conf->pool_thread_size > 0) { ret = hos_putobject_async(handle, request, data_len); } else { ret = hos_putobject_sync(handle, request, data_len); } return ret; } int hos_upload_file(const char *bucket, const char *file_path, put_finished_callback callback, void *userdata) { struct stat buffer; if (g_hos_handle.S3Client == NULL) { return HOS_INSTANCE_NOT_INIT; } if ((bucket == NULL) || (file_path == NULL)) { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_FATAL, __FUNCTION__, "error: bucket:%s, file_path:%s", bucket, file_path); return HOS_PARAMETER_ERROR; } if (stat(file_path, &buffer) == -1) { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_FATAL, __FUNCTION__, "error: The file:%s not exist", file_path); return HOS_FILE_NOT_EXIST; } return hos_upload_stream(bucket, file_path, NULL, buffer.st_size, callback, userdata); } int hos_upload_buf(const char *bucket, const char *object, const char *buf, size_t buf_len, put_finished_callback callback, void *userdata) { if (g_hos_handle.S3Client == NULL) { return HOS_INSTANCE_NOT_INIT; } if ((bucket == NULL) || (object == NULL) || (buf == NULL) || (buf_len == 0)) { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_FATAL, "hos_upload_buf", "bucket:%s, object:%s, buf:%s, buf_len:%lu", bucket, object, buf?"not null":"null", buf_len); return HOS_PARAMETER_ERROR; } return hos_upload_stream(bucket, object, buf, buf_len, callback, userdata); } int hos_open_handle(const char *bucket, const char *object, put_finished_callback callback, void *userdata, hos_handle handle) { if (g_hos_handle.S3Client == NULL) { return HOS_INSTANCE_NOT_INIT; } if ((bucket == NULL) || (object == NULL) || strlen(bucket) == 0 || strlen(object) == 0) { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_FATAL, __FUNCTION__, "error: bucket:%s, obejct:%s", (bucket == NULL)?"null":bucket, (object == NULL)?"null":object); return HOS_PARAMETER_ERROR; } handle = (hos_handle_t *)calloc(1, sizeof(hos_handle_t)); handle->bucket = (char *)calloc(1, strlen(bucket)+1); memcpy(handle->bucket, bucket, strlen(bucket)); handle->object = (char *)calloc(1, strlen(object)+1); memcpy(handle->object, object, strlen(object)); handle->callback = callback; handle->userdata = userdata; handle->handle_status = HOS_HANDLE_REGISTER; return HOS_CLIENT_OK; } int hos_write(hos_handle handle, const char *stream, size_t stream_len) { char num[128]; int ret = 0; data_info_t *data_info = NULL; hos_config_t *hos_conf = &g_hos_handle.hos_config; hos_func_thread_t *hos_func = &g_hos_handle.hos_func; size_t upload_len = 0; if (g_hos_handle.S3Client == NULL) { return HOS_INSTANCE_NOT_INIT; } if (stream == NULL) { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_FATAL, "hos_write", "error: stream:%s, stream_len:%lu.", stream?"not null":"null", stream_len); return HOS_PARAMETER_ERROR; } if (handle == NULL) { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_FATAL, __FUNCTION__, "error: fd info not find."); return HOS_HASH_NOT_FIND; } // create and configure the asynchronous put object request. Aws::S3::Model::PutObjectRequest request; //field_stat2 record if (hos_func->fs2_info.fs2_handle) { data_info = (data_info_t *)&hos_func->fs2_info.data_info; atomic_add(&data_info->rx_pkts, 1); atomic_add(&data_info->rx_bytes, stream_len); } if (handle->cache == NULL) { handle->cache = std::make_shared(); } Aws::String buffer(stream, stream_len); *handle->cache << buffer; handle->cache_rest -= stream_len; atomic_add(&data_info->cache, stream_len); if (handle->cache_count == 0 || --handle->cache_count) { //cache_count == 0,不设置cache_count的情况 //cache_count > 0,设置cache_count的情况 if (handle->cache_rest > 0) { return HOS_CLIENT_OK; } } request.SetBody(handle->cache); // add headers atomic_add(&(handle->position), 1); snprintf(num, 128, "%lu", atomic_read(&(handle->position))); Aws::Map headers; headers["x-hos-upload-type"] = "append"; headers["x-hos-position"] = num; request.SetMetadata(headers); handle->cache->seekg(0, std::ios_base::end); upload_len = handle->cache->tellg(); handle->cache->seekg(0, std::ios_base::beg); MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_DEBUG, __FUNCTION__, "debug: x-hos-posotion:%s", num); request.SetBucket(handle->bucket); request.SetKey(handle->object); if (hos_conf->pool_thread_size > 0) { ret = hos_putobject_async(handle, request, upload_len); } else { ret = hos_putobject_sync(handle, request, upload_len); } //恢复fd 的cache设置 if (handle->mode & APPEND_MODE) { atomic_sub(&data_info->cache, upload_len); handle->cache.reset(); handle->cache = NULL; handle->cache_rest = hos_conf->cache_size; handle->cache_count = hos_conf->cache_count; } return ret; } int hos_close_handle(hos_handle handle) { char num[128]; hos_config_t *hos_conf = &g_hos_handle.hos_config; size_t upload_len = 0; if (g_hos_handle.S3Client == NULL) { return HOS_INSTANCE_NOT_INIT; } if (handle == NULL) { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_FATAL, __FUNCTION__, "error: handle is null."); return HOS_PARAMETER_ERROR; } //close fd 之前发送append的缓存中内容 if ((handle->mode & BUFF_MODE) && (handle->mode & APPEND_MODE)) { if (handle->cache_rest != (long)hos_conf->cache_size && handle->cache != NULL) { // Create and configure the asynchronous put object request. Aws::S3::Model::PutObjectRequest request; request.SetBucket(handle->bucket); request.SetKey(handle->object); request.SetBody(handle->cache); // add headers atomic_add(&(handle->position), 1); snprintf(num, 128, "%lu", atomic_read(&(handle->position))); Aws::Map headers; headers["x-hos-upload-type"] = "append"; headers["x-hos-position"] = num; request.SetMetadata(headers); handle->cache->seekg(0, std::ios_base::end); upload_len = handle->cache->tellg(); handle->cache->seekg(0, std::ios_base::beg); if (hos_conf->pool_thread_size > 0) { hos_putobject_async(handle, request, upload_len); } else { hos_putobject_sync(handle, request, upload_len); } atomic_set(&g_hos_handle.hos_func.fs2_info.data_info.cache, 0); } } handle->handle_status = HOS_HANDLE_CANCEL; handle->cache.reset(); handle->cache = NULL; handle->cache_rest = hos_conf->cache_size; handle->cache_count = hos_conf->cache_count; if (hos_conf->pool_thread_size == 0) { //同步模式,立即释放fd hos_destroy_handle(handle); } else { //异步APPEND 模式,判断是否可以释放 //异步其他模式,在PutObjectAsyncFinished出释放fd if (handle->mode == (BUFF_MODE | APPEND_MODE) && handle->position == handle->recive_cnt) { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_DEBUG, __FUNCTION__, "debug: [%s:%s] upload completed. handle delete", handle->bucket, handle->object); hos_destroy_handle(handle); } } return HOS_CLIENT_OK; } int hos_shutdown_instance() { std::lock_guard locker(m_client_lock); size_t i = 0; hos_func_thread_t *hos_func = &g_hos_handle.hos_func; if (g_hos_handle.S3Client == NULL) { return HOS_INSTANCE_NOT_INIT; } if (g_hos_handle.count > 0 && --g_hos_handle.count) { MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_DEBUG, __FUNCTION__, "debug: hos client count:%lu.", g_hos_handle.count); return HOS_CLIENT_OK; } //先等待所有的task完成 while(1) { if (atomic_read(&g_hos_handle.task_num) == 0) break; usleep(500 * 1000); } Aws::Vector().swap(g_hos_handle.buckets); if (hos_func->fd_thread) { hos_func->fd_thread_status = 1; pthread_join(hos_func->fd_thread, NULL); } if (hos_func->fs2_thread) { hos_func->fs2_status = HOS_FS2_STOP; pthread_join(hos_func->fs2_thread, NULL); for (i = 0; i < FS2_RECORD_EVENTS; i++) { screen_stat_handle_t *fs2_handle = &hos_func->fs2_info.fs2_handle; if (*fs2_handle) { FS_stop(fs2_handle); *fs2_handle = NULL; } if (hos_func->fs2_info.line_ids) { free(hos_func->fs2_info.line_ids); hos_func->fs2_info.line_ids=NULL; } if (hos_func->fs2_info.column_ids) { free(hos_func->fs2_info.column_ids); hos_func->fs2_info.column_ids=NULL; } } } delete g_hos_handle.S3Client; g_hos_handle.S3Client = NULL; MESA_HANDLE_RUNTIME_LOG(g_hos_handle.log, RLOG_LV_DEBUG, __FUNCTION__, "debug: delete s3client."); Aws::ShutdownAPI(g_options); MESA_destroy_runtime_log_handle(g_hos_handle.log); g_hos_handle.log = NULL; memset(&g_hos_handle, 0 , sizeof(g_hos_handle)); if (g_hos_instance.hos_url_prefix) free((void *)g_hos_instance.hos_url_prefix); memset(&g_hos_instance, 0, sizeof(g_hos_instance)); return HOS_CLIENT_OK; }