client.py 26 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653
  1. #!/usr/bin/env python
  2. # -*- coding: utf-8 -*-
  3. # filename: client.py
  4. """
  5. Client module for Fastdfs 3.08
  6. author: scott yuan scottzer8@gmail.com
  7. date: 2012-06-21
  8. """
  9. from fdfs_client.tracker_client import *
  10. from fdfs_client.storage_client import *
  11. from fdfs_client.exceptions import *
  12. def get_tracker_conf(conf_path='client.conf'):
  13. cf = Fdfs_ConfigParser()
  14. tracker = {}
  15. try:
  16. cf.read(conf_path)
  17. timeout = cf.getint('__config__', 'connect_timeout')
  18. tracker_list = cf.get('__config__', 'tracker_server')
  19. if isinstance(tracker_list, str):
  20. tracker_list = [tracker_list]
  21. tracker_ip_list = []
  22. for tr in tracker_list:
  23. tracker_ip, tracker_port = tr.split(':')
  24. tracker_ip_list.append((tracker_ip, tracker_port))
  25. tracker['host_tuple'] = tuple(tracker_ip_list)
  26. tracker['timeout'] = timeout
  27. tracker['name'] = 'Tracker Pool'
  28. except:
  29. raise
  30. return tracker
  31. class Fdfs_client(object):
  32. """
  33. Class Fdfs_client implemented Fastdfs client protol ver 3.08.
  34. It's useful upload, download, delete file to or from fdfs server, etc. It's uses
  35. connection pool to manage connection to server.
  36. """
  37. def __init__(self, conf_path='/etc/fdfs/client.conf', poolclass=ConnectionPool):
  38. self.trackers = get_tracker_conf(conf_path)
  39. self.tracker_pool = poolclass(**self.trackers)
  40. self.timeout = self.trackers['timeout']
  41. self.storages = {}
  42. return None
  43. def __del__(self):
  44. try:
  45. self.pool.destroy()
  46. self.pool = None
  47. except:
  48. pass
  49. def get_storage(self, store_serv):
  50. store = self.storages.get((store_serv.ip_addr, store_serv.port), None)
  51. if store is None:
  52. store = Storage_client(store_serv.ip_addr, store_serv.port, self.timeout)
  53. self.storages[(store_serv.ip_addr, store_serv.port)] = store
  54. return store
  55. def get_store_serv(self, remote_file_id):
  56. '''
  57. Get store server info by remote_file_id.
  58. @author: LeoTse
  59. @param remote_file_id: string, file_id of file that is on storage server
  60. @return Storage_server object
  61. '''
  62. tmp = split_remote_fileid(remote_file_id)
  63. if not tmp:
  64. raise DataError('[-] Error: remote_file_id is invalid.(in delete file)')
  65. group_name, remote_filename = tmp
  66. tc = Tracker_client(self.tracker_pool)
  67. store_serv = tc.tracker_query_storage_update(group_name, remote_filename)
  68. return store_serv
  69. def upload_by_filename(self, filename, meta_dict = None):
  70. """
  71. Upload a file to Storage server.
  72. arguments:
  73. @filename: string, name of file that will be uploaded
  74. @meta_dict: dictionary e.g.:{
  75. 'ext_name' : 'jpg',
  76. 'file_size' : '10240B',
  77. 'width' : '160px',
  78. 'hight' : '80px'
  79. } meta_dict can be null
  80. @return dict {
  81. 'Group name' : group_name,
  82. 'Remote file_id' : remote_file_id,
  83. 'Status' : 'Upload successed.',
  84. 'Local file name' : local_file_name,
  85. 'Uploaded size' : upload_size,
  86. 'Storage IP' : storage_ip
  87. } if success else None
  88. """
  89. isfile, errmsg = fdfs_check_file(filename)
  90. if not isfile:
  91. raise DataError(errmsg + '(uploading)')
  92. tc = Tracker_client(self.tracker_pool)
  93. store_serv = tc.tracker_query_storage_stor_without_group()
  94. return self.get_storage(store_serv).storage_upload_by_filename(tc, store_serv, filename, meta_dict)
  95. def upload_by_file(self, filename, meta_dict=None):
  96. isfile, errmsg = fdfs_check_file(filename)
  97. if not isfile:
  98. raise DataError(errmsg + '(uploading)')
  99. tc = Tracker_client(self.tracker_pool)
  100. store_serv = tc.tracker_query_storage_stor_without_group()
  101. return self.get_storage(store_serv).storage_upload_by_file(tc, store_serv, filename, meta_dict)
  102. def upload_by_buffer(self, filebuffer, file_ext_name=None, meta_dict=None):
  103. """
  104. Upload a buffer to Storage server.
  105. arguments:
  106. @filebuffer: string, buffer
  107. @file_ext_name: string, file extend name
  108. @meta_dict: dictionary e.g.:{
  109. 'ext_name' : 'jpg',
  110. 'file_size' : '10240B',
  111. 'width' : '160px',
  112. 'hight' : '80px'
  113. }
  114. @return dict {
  115. 'Group name' : group_name,
  116. 'Remote file_id' : remote_file_id,
  117. 'Status' : 'Upload successed.',
  118. 'Local file name' : '',
  119. 'Uploaded size' : upload_size,
  120. 'Storage IP' : storage_ip
  121. } if success else None
  122. """
  123. if not filebuffer:
  124. raise DataError('[-] Error: argument filebuffer can not be null.')
  125. tc = Tracker_client(self.tracker_pool)
  126. store_serv = tc.tracker_query_storage_stor_without_group()
  127. return self.get_storage(store_serv).storage_upload_by_buffer(tc, store_serv, filebuffer, \
  128. file_ext_name, meta_dict)
  129. def upload_slave_by_filename(self, filename, remote_file_id, prefix_name, \
  130. meta_dict=None):
  131. """
  132. Upload slave file to Storage server.
  133. arguments:
  134. @filename: string, local file name
  135. @remote_file_id: string, remote file id
  136. @prefix_name: string
  137. @meta_dict: dictionary e.g.:{
  138. 'ext_name' : 'jpg',
  139. 'file_size' : '10240B',
  140. 'width' : '160px',
  141. 'hight' : '80px'
  142. }
  143. @return dictionary {
  144. 'Status' : 'Upload slave successed.',
  145. 'Local file name' : local_filename,
  146. 'Uploaded size' : upload_size,
  147. 'Remote file id' : remote_file_id,
  148. 'Storage IP' : storage_ip
  149. }
  150. """
  151. isfile, errmsg = fdfs_check_file(filename)
  152. if not isfile:
  153. raise DataError(errmsg + '(uploading slave)')
  154. tmp = split_remote_fileid(remote_file_id)
  155. if not tmp:
  156. raise DataError('[-] Error: remote_file_id is invalid.(uploading slave)')
  157. if not prefix_name:
  158. raise DataError('[-] Error: prefix_name can not be null.')
  159. group_name, remote_filename = tmp
  160. tc = Tracker_client(self.tracker_pool)
  161. store_serv = tc.tracker_query_storage_stor_with_group(group_name)
  162. store = self.get_storage(store_serv)
  163. try:
  164. ret_dict = store.storage_upload_slave_by_filename(tc, store_serv, filename, \
  165. prefix_name, remote_filename, \
  166. meta_dict=None)
  167. except:
  168. raise
  169. ret_dict['Status'] = 'Upload slave file successed.'
  170. return ret_dict
  171. def upload_slave_by_file(self, filename, remote_file_id, prefix_name, \
  172. meta_dict=None):
  173. """
  174. Upload slave file to Storage server.
  175. arguments:
  176. @filename: string, local file name
  177. @remote_file_id: string, remote file id
  178. @prefix_name: string
  179. @meta_dict: dictionary e.g.:{
  180. 'ext_name' : 'jpg',
  181. 'file_size' : '10240B',
  182. 'width' : '160px',
  183. 'hight' : '80px'
  184. }
  185. @return dictionary {
  186. 'Status' : 'Upload slave successed.',
  187. 'Local file name' : local_filename,
  188. 'Uploaded size' : upload_size,
  189. 'Remote file id' : remote_file_id,
  190. 'Storage IP' : storage_ip
  191. }
  192. """
  193. isfile, errmsg = fdfs_check_file(filename)
  194. if not isfile:
  195. raise DataError(errmsg + '(uploading slave)')
  196. tmp = split_remote_fileid(remote_file_id)
  197. if not tmp:
  198. raise DataError('[-] Error: remote_file_id is invalid.(uploading slave)')
  199. if not prefix_name:
  200. raise DataError('[-] Error: prefix_name can not be null.')
  201. group_name, remote_filename = tmp
  202. tc = Tracker_client(self.tracker_pool)
  203. store_serv = tc.tracker_query_storage_stor_with_group(group_name)
  204. store = self.get_storage(store_serv)
  205. try:
  206. ret_dict = store.storage_upload_slave_by_file(tc, store_serv, filename, \
  207. prefix_name, remote_filename, \
  208. meta_dict=None)
  209. except:
  210. raise
  211. ret_dict['Status'] = 'Upload slave file successed.'
  212. return ret_dict
  213. def upload_slave_by_buffer(self, filebuffer, remote_file_id, \
  214. meta_dict=None, file_ext_name=None):
  215. """
  216. Upload slave file by buffer
  217. arguments:
  218. @filebuffer: string
  219. @remote_file_id: string
  220. @meta_dict: dictionary e.g.:{
  221. 'ext_name' : 'jpg',
  222. 'file_size' : '10240B',
  223. 'width' : '160px',
  224. 'hight' : '80px'
  225. }
  226. @return dictionary {
  227. 'Status' : 'Upload slave successed.',
  228. 'Local file name' : local_filename,
  229. 'Uploaded size' : upload_size,
  230. 'Remote file id' : remote_file_id,
  231. 'Storage IP' : storage_ip
  232. }
  233. """
  234. if not filebuffer:
  235. raise DataError('[-] Error: argument filebuffer can not be null.')
  236. tmp = split_remote_fileid(remote_file_id)
  237. if not tmp:
  238. raise DataError('[-] Error: remote_file_id is invalid.(uploading slave)')
  239. group_name, remote_filename = tmp
  240. tc = Tracker_client(self.tracker_pool)
  241. store_serv = tc.tracker_query_storage_update(group_name, remote_filename)
  242. store = self.get_storage(store_serv)
  243. return store.storage_upload_slave_by_buffer(tc, store_serv, filebuffer, \
  244. remote_filename, meta_dict, \
  245. file_ext_name)
  246. def upload_appender_by_filename(self, local_filename, meta_dict=None):
  247. """
  248. Upload an appender file by filename.
  249. arguments:
  250. @local_filename: string
  251. @meta_dict: dictionary e.g.:{
  252. 'ext_name' : 'jpg',
  253. 'file_size' : '10240B',
  254. 'width' : '160px',
  255. 'hight' : '80px'
  256. } Notice: it can be null
  257. @return dict {
  258. 'Group name' : group_name,
  259. 'Remote file_id' : remote_file_id,
  260. 'Status' : 'Upload successed.',
  261. 'Local file name' : '',
  262. 'Uploaded size' : upload_size,
  263. 'Storage IP' : storage_ip
  264. } if success else None
  265. """
  266. isfile, errmsg = fdfs_check_file(local_filename)
  267. if not isfile:
  268. raise DataError(errmsg + '(uploading appender)')
  269. tc = Tracker_client(self.tracker_pool)
  270. store_serv = tc.tracker_query_storage_stor_without_group()
  271. store = self.get_storage(store_serv)
  272. return store.storage_upload_appender_by_filename(tc, store_serv, \
  273. local_filename, meta_dict)
  274. def upload_appender_by_file(self, local_filename, meta_dict=None):
  275. """
  276. Upload an appender file by file.
  277. arguments:
  278. @local_filename: string
  279. @meta_dict: dictionary e.g.:{
  280. 'ext_name' : 'jpg',
  281. 'file_size' : '10240B',
  282. 'width' : '160px',
  283. 'hight' : '80px'
  284. } Notice: it can be null
  285. @return dict {
  286. 'Group name' : group_name,
  287. 'Remote file_id' : remote_file_id,
  288. 'Status' : 'Upload successed.',
  289. 'Local file name' : '',
  290. 'Uploaded size' : upload_size,
  291. 'Storage IP' : storage_ip
  292. } if success else None
  293. """
  294. isfile, errmsg = fdfs_check_file(local_filename)
  295. if not isfile:
  296. raise DataError(errmsg + '(uploading appender)')
  297. tc = Tracker_client(self.tracker_pool)
  298. store_serv = tc.tracker_query_storage_stor_without_group()
  299. store = self.get_storage(store_serv)
  300. return store.storage_upload_appender_by_file(tc, store_serv, \
  301. local_filename, meta_dict)
  302. def upload_appender_by_buffer(self, filebuffer, file_ext_name=None, meta_dict=None):
  303. """
  304. Upload a buffer to Storage server.
  305. arguments:
  306. @filebuffer: string
  307. @file_ext_name: string, can be null
  308. @meta_dict: dictionary, can be null
  309. @return dict {
  310. 'Group name' : group_name,
  311. 'Remote file_id' : remote_file_id,
  312. 'Status' : 'Upload successed.',
  313. 'Local file name' : '',
  314. 'Uploaded size' : upload_size,
  315. 'Storage IP' : storage_ip
  316. } if success else None
  317. """
  318. if not filebuffer:
  319. raise DataError('[-] Error: argument filebuffer can not be null.')
  320. tc = Tracker_client(self.tracker_pool)
  321. store_serv = tc.tracker_query_storage_stor_without_group()
  322. store = self.get_storage(store_serv)
  323. return store.storage_upload_appender_by_buffer(tc, store_serv, \
  324. filebuffer, meta_dict, \
  325. file_ext_name)
  326. def delete_file(self, remote_file_id):
  327. """
  328. Delete a file from Storage server.
  329. arguments:
  330. @remote_file_id: string, file_id of file that is on storage server
  331. @return tuple ('Delete file successed.', remote_file_id, storage_ip)
  332. """
  333. tmp = split_remote_fileid(remote_file_id)
  334. if not tmp:
  335. raise DataError('[-] Error: remote_file_id is invalid.(in delete file)')
  336. group_name, remote_filename = tmp
  337. tc = Tracker_client(self.tracker_pool)
  338. store_serv = tc.tracker_query_storage_update(group_name, remote_filename)
  339. store = self.get_storage(store_serv)
  340. return store.storage_delete_file(tc, store_serv, remote_filename)
  341. def download_to_file(self, local_filename, remote_file_id, offset=0, down_bytes=0):
  342. """
  343. Download a file from Storage server.
  344. arguments:
  345. @local_filename: string, local name of file
  346. @remote_file_id: string, file_id of file that is on storage server
  347. @offset: long
  348. @downbytes: long
  349. @return dict {
  350. 'Remote file_id' : remote_file_id,
  351. 'Content' : local_filename,
  352. 'Download size' : downloaded_size,
  353. 'Storage IP' : storage_ip
  354. }
  355. """
  356. tmp = split_remote_fileid(remote_file_id)
  357. if not tmp:
  358. raise DataError('[-] Error: remote_file_id is invalid.(in download file)')
  359. group_name, remote_filename = tmp
  360. if not offset:
  361. file_offset = offset
  362. if not down_bytes:
  363. download_bytes = down_bytes
  364. tc = Tracker_client(self.tracker_pool)
  365. store_serv = tc.tracker_query_storage_fetch(group_name, remote_filename)
  366. store = Storage_client(store_serv.ip_addr, store_serv.port, self.timeout)
  367. return store.storage_download_to_file(tc, store_serv, local_filename, file_offset, download_bytes,
  368. remote_filename)
  369. def download_to_buffer(self, remote_file_id, offset=0, down_bytes=0):
  370. """
  371. Download a file from Storage server and store in buffer.
  372. arguments:
  373. @remote_file_id: string, file_id of file that is on storage server
  374. @offset: long
  375. @down_bytes: long
  376. @return dict {
  377. 'Remote file_id' : remote_file_id,
  378. 'Content' : file_buffer,
  379. 'Download size' : downloaded_size,
  380. 'Storage IP' : storage_ip
  381. }
  382. """
  383. tmp = split_remote_fileid(remote_file_id)
  384. if not tmp:
  385. raise DataError('[-] Error: remote_file_id is invalid.(in download file)')
  386. group_name, remote_filename = tmp
  387. if not offset:
  388. file_offset = offset
  389. if not down_bytes:
  390. download_bytes = down_bytes
  391. tc = Tracker_client(self.tracker_pool)
  392. store_serv = tc.tracker_query_storage_fetch(group_name, remote_filename)
  393. store = Storage_client(store_serv.ip_addr, store_serv.port, self.timeout)
  394. file_buffer = None
  395. return store.storage_download_to_buffer(tc, store_serv, file_buffer, \
  396. file_offset, download_bytes, \
  397. remote_filename)
  398. def list_one_group(self, group_name):
  399. """
  400. List one group information.
  401. arguments:
  402. @group_name: string, group name will be list
  403. @return Group_info, instance
  404. """
  405. tc = Tracker_client(self.tracker_pool)
  406. return tc.tracker_list_one_group(group_name)
  407. def list_servers(self, group_name, storage_ip=None):
  408. """
  409. List all storage servers information in a group
  410. arguments:
  411. @group_name: string
  412. @return dictionary {
  413. 'Group name' : group_name,
  414. 'Servers' : server list,
  415. }
  416. """
  417. tc = Tracker_client(self.tracker_pool)
  418. return tc.tracker_list_servers(group_name, storage_ip)
  419. def list_all_groups(self):
  420. """
  421. List all group information.
  422. @return dictionary {
  423. 'Groups count' : group_count,
  424. 'Groups' : list of groups
  425. }
  426. """
  427. tc = Tracker_client(self.tracker_pool)
  428. return tc.tracker_list_all_groups()
  429. def get_meta_data(self, remote_file_id):
  430. """
  431. Get meta data of remote file.
  432. arguments:
  433. @remote_fileid: string, remote file id
  434. @return dictionary, meta data
  435. """
  436. tmp = split_remote_fileid(remote_file_id)
  437. if not tmp:
  438. raise DataError('[-] Error: remote_file_id is invalid.(in get meta data)')
  439. group_name, remote_filename = tmp
  440. tc = Tracker_client(self.tracker_pool)
  441. store_serv = tc.tracker_query_storage_update(group_name, remote_filename)
  442. store = Storage_client(store_serv.ip_addr, store_serv.port, self.timeout)
  443. return store.storage_get_metadata(tc, store_serv, remote_filename)
  444. def set_meta_data(self, remote_file_id, meta_dict, op_flag=STORAGE_SET_METADATA_FLAG_OVERWRITE):
  445. """
  446. Set meta data of remote file.
  447. arguments:
  448. @remote_file_id: string
  449. @meta_dict: dictionary
  450. @op_flag: char, 'O' for overwrite, 'M' for merge
  451. @return dictionary {
  452. 'Status' : status,
  453. 'Storage IP' : storage_ip
  454. }
  455. """
  456. tmp = split_remote_fileid(remote_file_id)
  457. if not tmp:
  458. raise DataError('[-] Error: remote_file_id is invalid.(in set meta data)')
  459. group_name, remote_filename = tmp
  460. tc = Tracker_client(self.tracker_pool)
  461. try:
  462. store_serv = tc.tracker_query_storage_update(group_name, remote_filename)
  463. store = Storage_client(store_serv.ip_addr, store_serv.port, self.timeout)
  464. status = store.storage_set_metadata(tc, store_serv, remote_filename, meta_dict)
  465. except (ConnectionError, ResponseError, DataError):
  466. raise
  467. # if status == 2:
  468. # raise DataError('[-] Error: remote file %s is not exist.' % remote_file_id)
  469. if status != 0:
  470. raise DataError('[-] Error: %d, %s' % (status, os.strerror(status)))
  471. ret_dict = {'Status': 'Set meta data success.', 'Storage IP': store_serv.ip_addr}
  472. return ret_dict
  473. def append_by_filename(self, local_filename, remote_fileid):
  474. isfile, errmsg = fdfs_check_file(local_filename)
  475. if not isfile:
  476. raise DataError(errmsg + '(append)')
  477. tmp = split_remote_fileid(remote_fileid)
  478. if not tmp:
  479. raise DataError('[-] Error: remote_file_id is invalid.(append)')
  480. group_name, appended_filename = tmp
  481. tc = Tracker_client(self.tracker_pool)
  482. store_serv = tc.tracker_query_storage_update(group_name, appended_filename)
  483. store = Storage_client(store_serv.ip_addr, store_serv.port, self.timeout)
  484. return store.storage_append_by_filename(tc, store_serv, local_filename, \
  485. appended_filename)
  486. def append_by_file(self, local_filename, remote_fileid):
  487. isfile, errmsg = fdfs_check_file(local_filename)
  488. if not isfile:
  489. raise DataError(errmsg + '(append)')
  490. tmp = split_remote_fileid(remote_fileid)
  491. if not tmp:
  492. raise DataError('[-] Error: remote_file_id is invalid.(append)')
  493. group_name, appended_filename = tmp
  494. tc = Tracker_client(self.tracker_pool)
  495. store_serv = tc.tracker_query_storage_update(group_name, appended_filename)
  496. store = Storage_client(store_serv.ip_addr, store_serv.port, self.timeout)
  497. return store.storage_append_by_file(tc, store_serv, local_filename, \
  498. appended_filename)
  499. def append_by_buffer(self, file_buffer, remote_fileid):
  500. if not file_buffer:
  501. raise DataError('[-] Error: file_buffer can not be null.')
  502. tmp = split_remote_fileid(remote_fileid)
  503. if not tmp:
  504. raise DataError('[-] Error: remote_file_id is invalid.(append)')
  505. group_name, appended_filename = tmp
  506. tc = Tracker_client(self.tracker_pool)
  507. store_serv = tc.tracker_query_storage_update(group_name, appended_filename)
  508. store = Storage_client(store_serv.ip_addr, store_serv.port, self.timeout)
  509. return store.storage_append_by_buffer(tc, store_serv, file_buffer, \
  510. appended_filename)
  511. def truncate_file(self, truncated_filesize, appender_fileid):
  512. """
  513. Truncate file in Storage server.
  514. arguments:
  515. @truncated_filesize: long
  516. @appender_fileid: remote_fileid
  517. @return: dictionary {
  518. 'Status' : 'Truncate successed.',
  519. 'Storage IP' : storage_ip
  520. }
  521. """
  522. trunc_filesize = truncated_filesize
  523. tmp = split_remote_fileid(appender_fileid)
  524. if not tmp:
  525. raise DataError('[-] Error: appender_fileid is invalid.(truncate)')
  526. group_name, appender_filename = tmp
  527. tc = Tracker_client(self.tracker_pool)
  528. store_serv = tc.tracker_query_storage_update(group_name, appender_filename)
  529. store = Storage_client(store_serv.ip_addr, store_serv.port, self.timeout)
  530. return store.storage_truncate_file(tc, store_serv, trunc_filesize, \
  531. appender_filename)
  532. def modify_by_filename(self, filename, appender_fileid, offset=0):
  533. """
  534. Modify a file in Storage server by file.
  535. arguments:
  536. @filename: string, local file name
  537. @offset: long, file offset
  538. @appender_fileid: string, remote file id
  539. @return: dictionary {
  540. 'Status' : 'Modify successed.',
  541. 'Storage IP' : storage_ip
  542. }
  543. """
  544. isfile, errmsg = fdfs_check_file(filename)
  545. if not isfile:
  546. raise DataError(errmsg + '(modify)')
  547. filesize = os.stat(filename).st_size
  548. tmp = split_remote_fileid(appender_fileid)
  549. if not tmp:
  550. raise DataError('[-] Error: remote_fileid is invalid.(modify)')
  551. group_name, appender_filename = tmp
  552. if not offset:
  553. file_offset = offset
  554. else:
  555. file_offset = 0
  556. tc = Tracker_client(self.tracker_pool)
  557. store_serv = tc.tracker_query_storage_update(group_name, appender_filename)
  558. store = Storage_client(store_serv.ip_addr, store_serv.port, self.timeout)
  559. return store.storage_modify_by_filename(tc, store_serv, filename, file_offset, \
  560. filesize, appender_filename)
  561. def modify_by_file(self, filename, appender_fileid, offset=0):
  562. """
  563. Modify a file in Storage server by file.
  564. arguments:
  565. @filename: string, local file name
  566. @offset: long, file offset
  567. @appender_fileid: string, remote file id
  568. @return: dictionary {
  569. 'Status' : 'Modify successed.',
  570. 'Storage IP' : storage_ip
  571. }
  572. """
  573. isfile, errmsg = fdfs_check_file(filename)
  574. if not isfile:
  575. raise DataError(errmsg + '(modify)')
  576. filesize = os.stat(filename).st_size
  577. tmp = split_remote_fileid(appender_fileid)
  578. if not tmp:
  579. raise DataError('[-] Error: remote_fileid is invalid.(modify)')
  580. group_name, appender_filename = tmp
  581. if not offset:
  582. file_offset = offset
  583. else:
  584. file_offset = 0
  585. tc = Tracker_client(self.tracker_pool)
  586. store_serv = tc.tracker_query_storage_update(group_name, appender_filename)
  587. store = Storage_client(store_serv.ip_addr, store_serv.port, self.timeout)
  588. return store.storage_modify_by_file(tc, store_serv, filename, file_offset, \
  589. filesize, appender_filename)
  590. def modify_by_buffer(self, filebuffer, appender_fileid, offset=0):
  591. """
  592. Modify a file in Storage server by buffer.
  593. arguments:
  594. @filebuffer: string, file buffer
  595. @offset: long, file offset
  596. @appender_fileid: string, remote file id
  597. @return: dictionary {
  598. 'Status' : 'Modify successed.',
  599. 'Storage IP' : storage_ip
  600. }
  601. """
  602. if not filebuffer:
  603. raise DataError('[-] Error: filebuffer can not be null.(modify)')
  604. filesize = len(filebuffer)
  605. tmp = split_remote_fileid(appender_fileid)
  606. if not tmp:
  607. raise DataError('[-] Error: remote_fileid is invalid.(modify)')
  608. group_name, appender_filename = tmp
  609. if not offset:
  610. file_offset = offset
  611. else:
  612. file_offset = 0
  613. tc = Tracker_client(self.tracker_pool)
  614. store_serv = tc.tracker_query_storage_update(group_name, appender_filename)
  615. store = Storage_client(store_serv.ip_addr, store_serv.port, self.timeout)
  616. return store.storage_modify_by_buffer(tc, store_serv, filebuffer, file_offset, \
  617. filesize, appender_filename)