文章目录

  • 1.fastfds文件服务器搭建
    • 1.1环境准备
    • 1.2环境编译
    • 1.3磁盘目录
    • 1.4安装libfascommon
    • 1.5安装FastDFS
    • 1.6安装fastdfs-nginx-module
    • 1.7安装nginx
    • 1.8配置文件
      • 1.8.1 tracker配置
      • 1.8.2 storage配置
      • 1.8.3配置nginx
  • 2启动
    • 2.1关闭防火墙
  • 2.2启动tracker
  • 2.3启动storage
  • 2.4启动nginx
  • 2.4检测
  • 3.java fastDFS cilent

1.fastfds文件服务器搭建

1.1环境准备

名称 说明
centos 7x
libfatscommon FastDFS分离出的一些公用函数包
FastDFS FastDFS本体
fastdfs-nginx-module FastDFS和nginx的关联模块
nginx nginx1.15.4

1.2环境编译

yum install git gcc gcc-c++ make automake autoconf libtool pcre pcre-devel zlib zlib-devel openssl-devel wget vim -y

1.3磁盘目录

说明 位置
安装包存放位置 usr/local/src
数据存储位置 /home/dfs/resources
日志存放位置 /home/dfs/log
mkdir /home/dfs/resources #创建数据存储目录
mkdir /home/dfs/log #创建日志存储目录

1.4安装libfascommon

git clone https://github.com/happyfish100/libfastcommon.git --depth 1
cd libfastcommon/
./make.sh && ./make.sh install #编译安装

1.5安装FastDFS

cd ../ #返回上一级目录
git clone https://github.com/happyfish100/fastdfs.git --depth 1
cd fastdfs/
./make.sh && ./make.sh install #编译安装
#配置文件准备
cp /etc/fdfs/tracker.conf.sample /etc/fdfs/tracker.conf
cp /etc/fdfs/storage.conf.sample /etc/fdfs/storage.conf
cp /etc/fdfs/client.conf.sample /etc/fdfs/client.conf #客户端文件,测试用
cp /usr/local/src/fastdfs/conf/http.conf /etc/fdfs/ #供nginx访问使用
cp /usr/local/src/fastdfs/conf/mime.types /etc/fdfs/ #供nginx访问使用

1.6安装fastdfs-nginx-module

cd ../ #返回上一级目录
git clone https://github.com/happyfish100/fastdfs-nginx-module.git --depth 1
cp /usr/local/src/fastdfs-nginx-module/src/mod_fastdfs.conf /etc/fdfs

1.7安装nginx

wget http://nginx.org/download/nginx-1.15.4.tar.gz #下载nginx压缩包
tar -zxvf nginx-1.15.4.tar.gz #解压
cd nginx-1.15.4/
#添加fastdfs-nginx-module模块
./configure --add-module=/usr/local/src/fastdfs-nginx-module/src/
make && make install #编译安装

1.8配置文件

1.8.1 tracker配置

/etc/fdfs/tracker.conf
修改:
bind_addr =192.168.177.128 //服务器IP ,IP根据本机IP来填写
base_path = /home/dfs/log //日志存储位置

# is this config file disabled
# false for enabled
# true for disabled
disabled = false #当前配置是否可用 false:可用 true:不可用# bind an address of this host
# empty for bind all addresses of this host
bind_addr =192.168.177.128 # the tracker server port
port = 22122# connect timeout in seconds
# default value is 30
# Note: in the intranet network (LAN), 2 seconds is enough.
connect_timeout = 5# network timeout in seconds for send and recv
# default value is 30
network_timeout = 60# the base path to store data and log files
base_path = /home/dfs/log# max concurrent connections this server support
# you should set this parameter larger, eg. 10240
# default value is 256
max_connections = 1024# accept thread count
# default value is 1 which is recommended
# since V4.07
accept_threads = 1# work thread count
# work threads to deal network io
# default value is 4
# since V2.00
work_threads = 4# the min network buff size
# default value 8KB
min_buff_size = 8KB# the max network buff size
# default value 128KB
max_buff_size = 128KB# the method for selecting group to upload files
# 0: round robin
# 1: specify group
# 2: load balance, select the max free space group to upload file
store_lookup = 2 #上传组(卷)的方式:0轮询,1:指定,2:负载平衡(剩余空间最大选择)# which group to upload file
# when store_lookup set to 1, must set store_group to the group name
store_group = group2# which storage server to upload file
# 0: round robin (default)
# 1: the first server order by ip address
# 2: the first server order by priority (the minimal)
# Note: if use_trunk_file set to true, must set store_server to 1 or 2
store_server = 0# which path (means disk or mount point) of the storage server to upload file
# 0: round robin
# 2: load balance, select the max free space path to upload file
store_path = 0# which storage server to download file
# 0: round robin (default)
# 1: the source storage server which the current file uploaded to
download_server = 0# reserved storage space for system or other applications.
# if the free(available) space of any stoarge server in
# a group <= reserved_storage_space, no file can be uploaded to this group.
# bytes unit can be one of follows:
### G or g for gigabyte(GB)
### M or m for megabyte(MB)
### K or k for kilobyte(KB)
### no unit for byte(B)
### XX.XX% as ratio such as: reserved_storage_space = 10%
reserved_storage_space = 20%#standard log level as syslog, case insensitive, value list:
### emerg for emergency
### alert
### crit for critical
### error
### warn for warning
### notice
### info
### debug
log_level = info#unix group name to run this program,
#not set (empty) means run by the group of current user
run_by_group=#unix username to run this program,
#not set (empty) means run by current user
run_by_user =# allow_hosts can ocur more than once, host can be hostname or ip address,
# "*" (only one asterisk) means match all ip addresses
# we can use CIDR ips like 192.168.5.64/26
# and also use range like these: 10.0.1.[0-254] and host[01-08,20-25].domain.com
# for example:
# allow_hosts=10.0.1.[1-15,20]
# allow_hosts=host[01-08,20-25].domain.com
# allow_hosts=192.168.5.64/26
allow_hosts = *# sync log buff to disk every interval seconds
# default value is 10 seconds
sync_log_buff_interval = 1# check storage server alive interval seconds
check_active_interval = 120# thread stack size, should >= 64KB
# default value is 256KB
thread_stack_size = 256KB# auto adjust when the ip address of the storage server changed
# default value is true
storage_ip_changed_auto_adjust = true# storage sync file max delay seconds
# default value is 86400 seconds (one day)
# since V2.00
storage_sync_file_max_delay = 86400# the max time of storage sync a file
# default value is 300 seconds
# since V2.00
storage_sync_file_max_time = 300# if use a trunk file to store several small files
# default value is false
# since V3.00
use_trunk_file = false # the min slot size, should <= 4KB
# default value is 256 bytes
# since V3.00
slot_min_size = 256# the max slot size, should > slot_min_size
# store the upload file to trunk file when it's size <=  this value
# default value is 16MB
# since V3.00
slot_max_size = 1MB# the alignment size to allocate the trunk space
# default value is 0 (never align)
# since V6.05
# NOTE: the larger the alignment size, the less likely of disk
#       fragmentation, but the more space is wasted.
trunk_alloc_alignment_size = 256# if merge contiguous free spaces of trunk file
# default value is false
# since V6.05
trunk_free_space_merge = true# if delete / reclaim the unused trunk files
# default value is false
# since V6.05
delete_unused_trunk_files = false# the trunk file size, should >= 4MB
# default value is 64MB
# since V3.00
trunk_file_size = 64MB# if create trunk file advancely
# default value is false
# since V3.06
trunk_create_file_advance = false# the time base to create trunk file
# the time format: HH:MM
# default value is 02:00
# since V3.06
trunk_create_file_time_base = 02:00# the interval of create trunk file, unit: second
# default value is 38400 (one day)
# since V3.06
trunk_create_file_interval = 86400# the threshold to create trunk file
# when the free trunk file size less than the threshold,
# will create he trunk files
# default value is 0
# since V3.06
trunk_create_file_space_threshold = 20G# if check trunk space occupying when loading trunk free spaces
# the occupied spaces will be ignored
# default value is false
# since V3.09
# NOTICE: set this parameter to true will slow the loading of trunk spaces
# when startup. you should set this parameter to true when neccessary.
trunk_init_check_occupying = false# if ignore storage_trunk.dat, reload from trunk binlog
# default value is false
# since V3.10
# set to true once for version upgrade when your version less than V3.10
trunk_init_reload_from_binlog = false# the min interval for compressing the trunk binlog file
# unit: second, 0 means never compress
# FastDFS compress the trunk binlog when trunk init and trunk destroy
# recommand to set this parameter to 86400 (one day)
# default value is 0
# since V5.01
trunk_compress_binlog_min_interval = 86400# the interval for compressing the trunk binlog file
# unit: second, 0 means never compress
# recommand to set this parameter to 86400 (one day)
# default value is 0
# since V6.05
trunk_compress_binlog_interval = 86400# compress the trunk binlog time base, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 03:00
# since V6.05
trunk_compress_binlog_time_base = 03:00# max backups for the trunk binlog file
# default value is 0 (never backup)
# since V6.05
trunk_binlog_max_backups = 7# if use storage server ID instead of IP address
# if you want to use dual IPs for storage server, you MUST set
# this parameter to true, and configure the dual IPs in the file
# configured by following item "storage_ids_filename", such as storage_ids.conf
# default value is false
# since V4.00
use_storage_id = false# specify storage ids filename, can use relative or absolute path
# this parameter is valid only when use_storage_id set to true
# since V4.00
storage_ids_filename = storage_ids.conf# id type of the storage server in the filename, values are:
## ip: the ip address of the storage server
## id: the server id of the storage server
# this paramter is valid only when use_storage_id set to true
# default value is ip
# since V4.03
id_type_in_filename = id# if store slave file use symbol link
# default value is false
# since V4.01
store_slave_file_use_link = false# if rotate the error log every day
# default value is false
# since V4.02
rotate_error_log = false# rotate error log time base, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 00:00
# since V4.02
error_log_rotate_time = 00:00# if compress the old error log by gzip
# default value is false
# since V6.04
compress_old_error_log = false# compress the error log days before
# default value is 1
# since V6.04
compress_error_log_days_before = 7# rotate error log when the log file exceeds this size
# 0 means never rotates log file by log file size
# default value is 0
# since V4.02
rotate_error_log_size = 0# keep days of the log files
# 0 means do not delete old log files
# default value is 0
log_file_keep_days = 0# if use connection pool
# default value is false
# since V4.05
use_connection_pool = true# connections whose the idle time exceeds this time will be closed
# unit: second
# default value is 3600
# since V4.05
connection_pool_max_idle_time = 3600# HTTP port on this tracker server
http.server_port = 8080# check storage HTTP server alive interval seconds
# <= 0 for never check
# default value is 30
http.check_alive_interval = 30# check storage HTTP server alive type, values are:
#   tcp : connect to the storge server with HTTP port only,
#        do not request and get response
#   http: storage check alive url must return http status 200
# default value is tcp
http.check_alive_type = tcp# check storage HTTP server alive uri/url
# NOTE: storage embed HTTP server support uri: /status.html
http.check_alive_uri = /status.html

1.8.2 storage配置

/etc/fdfs/storage.conf
修改:
bind_addr =192.168.177.128
base_path = /home/dfs/log
store_path0 = /home/dfs/resources
tracker_server = 192.168.177.128:22122 tracker服务器IP和端口
http.server_port=80 # http访问文件的端口(默认8888,看情况修改,和nginx中保持一致)

# is this config file disabled
# false for enabled
# true for disabled
disabled = false# the name of the group this storage server belongs to
#
# comment or remove this item for fetching from tracker server,
# in this case, use_storage_id must set to true in tracker.conf,
# and storage_ids.conf must be configured correctly.
group_name = group1# bind an address of this host
# empty for bind all addresses of this host
bind_addr =192.168.177.128# if bind an address of this host when connect to other servers
# (this storage server as a client)
# true for binding the address configured by the above parameter: "bind_addr"
# false for binding any address of this host
client_bind = true# the storage server port
port = 23000# connect timeout in seconds
# default value is 30
# Note: in the intranet network (LAN), 2 seconds is enough.
connect_timeout = 5# network timeout in seconds for send and recv
# default value is 30
network_timeout = 60# the heart beat interval in seconds
# the storage server send heartbeat to tracker server periodically
# default value is 30
heart_beat_interval = 30# disk usage report interval in seconds
# the storage server send disk usage report to tracker server periodically
# default value is 300
stat_report_interval = 60# the base path to store data and log files
# NOTE: the binlog files maybe are large, make sure
#       the base path has enough disk space,
#       eg. the disk free space should > 50GB
base_path = /home/dfs/log# max concurrent connections the server supported,
# you should set this parameter larger, eg. 10240
# default value is 256
max_connections = 1024# the buff size to recv / send data from/to network
# this parameter must more than 8KB
# 256KB or 512KB is recommended
# default value is 64KB
# since V2.00
buff_size = 256KB# accept thread count
# default value is 1 which is recommended
# since V4.07
accept_threads = 1# work thread count
# work threads to deal network io
# default value is 4
# since V2.00
work_threads = 4# if disk read / write separated
##  false for mixed read and write
##  true for separated read and write
# default value is true
# since V2.00
disk_rw_separated = true# disk reader thread count per store path
# for mixed read / write, this parameter can be 0
# default value is 1
# since V2.00
disk_reader_threads = 1# disk writer thread count per store path
# for mixed read / write, this parameter can be 0
# default value is 1
# since V2.00
disk_writer_threads = 1# when no entry to sync, try read binlog again after X milliseconds
# must > 0, default value is 200ms
sync_wait_msec = 50# after sync a file, usleep milliseconds
# 0 for sync successively (never call usleep)
sync_interval = 0# storage sync start time of a day, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
sync_start_time = 00:00# storage sync end time of a day, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
sync_end_time = 23:59# write to the mark file after sync N files
# default value is 500
write_mark_file_freq = 500# disk recovery thread count
# default value is 1
# since V6.04
disk_recovery_threads = 3# store path (disk or mount point) count, default value is 1
store_path_count = 1# store_path#, based on 0, to configure the store paths to store files
# if store_path0 not exists, it's value is base_path (NOT recommended)
# the paths must be exist.
#
# IMPORTANT NOTE:
#       the store paths' order is very important, don't mess up!!!
#       the base_path should be independent (different) of the store pathsstore_path0 = /home/dfs/resources
#store_path1 = /home/yuqing/fastdfs2# subdir_count  * subdir_count directories will be auto created under each
# store_path (disk), value can be 1 to 256, default value is 256
subdir_count_per_path = 256# tracker_server can ocur more than once for multi tracker servers.
# the value format of tracker_server is "HOST:PORT",
#   the HOST can be hostname or ip address,
#   and the HOST can be dual IPs or hostnames seperated by comma,
#   the dual IPS must be an inner (intranet) IP and an outer (extranet) IP,
#   or two different types of inner (intranet) IPs.
#   for example: 192.168.2.100,122.244.141.46:22122
#   another eg.: 192.168.1.10,172.17.4.21:22122tracker_server = 192.168.177.128:22122
#tracker_server = 192.168.209.122:22122#standard log level as syslog, case insensitive, value list:
### emerg for emergency
### alert
### crit for critical
### error
### warn for warning
### notice
### info
### debug
log_level = info#unix group name to run this program,
#not set (empty) means run by the group of current user
run_by_group =#unix username to run this program,
#not set (empty) means run by current user
run_by_user =# allow_hosts can ocur more than once, host can be hostname or ip address,
# "*" (only one asterisk) means match all ip addresses
# we can use CIDR ips like 192.168.5.64/26
# and also use range like these: 10.0.1.[0-254] and host[01-08,20-25].domain.com
# for example:
# allow_hosts=10.0.1.[1-15,20]
# allow_hosts=host[01-08,20-25].domain.com
# allow_hosts=192.168.5.64/26
allow_hosts = *# the mode of the files distributed to the data path
# 0: round robin(default)
# 1: random, distributted by hash code
file_distribute_path_mode = 0# valid when file_distribute_to_path is set to 0 (round robin).
# when the written file count reaches this number, then rotate to next path.
# rotate to the first path (00/00) after the last path (such as FF/FF).
# default value is 100
file_distribute_rotate_count = 100# call fsync to disk when write big file
# 0: never call fsync
# other: call fsync when written bytes >= this bytes
# default value is 0 (never call fsync)
fsync_after_written_bytes = 0# sync log buff to disk every interval seconds
# must > 0, default value is 10 seconds
sync_log_buff_interval = 1# sync binlog buff / cache to disk every interval seconds
# default value is 60 seconds
sync_binlog_buff_interval = 1# sync storage stat info to disk every interval seconds
# default value is 300 seconds
sync_stat_file_interval = 300# thread stack size, should >= 512KB
# default value is 512KB
thread_stack_size = 512KB# the priority as a source server for uploading file.
# the lower this value, the higher its uploading priority.
# default value is 10
upload_priority = 10# the NIC alias prefix, such as eth in Linux, you can see it by ifconfig -a
# multi aliases split by comma. empty value means auto set by OS type
# default values is empty
if_alias_prefix =# if check file duplicate, when set to true, use FastDHT to store file indexes
# 1 or yes: need check
# 0 or no: do not check
# default value is 0
check_file_duplicate = 0# file signature method for check file duplicate
## hash: four 32 bits hash code
## md5: MD5 signature
# default value is hash
# since V4.01
file_signature_method = hash# namespace for storing file indexes (key-value pairs)
# this item must be set when check_file_duplicate is true / on
key_namespace = FastDFS# set keep_alive to 1 to enable persistent connection with FastDHT servers
# default value is 0 (short connection)
keep_alive = 0# you can use "#include filename" (not include double quotes) directive to
# load FastDHT server list, when the filename is a relative path such as
# pure filename, the base path is the base path of current/this config file.
# must set FastDHT server list when check_file_duplicate is true / on
# please see INSTALL of FastDHT for detail
##include /home/yuqing/fastdht/conf/fdht_servers.conf# if log to access log
# default value is false
# since V4.00
use_access_log = false# if rotate the access log every day
# default value is false
# since V4.00
rotate_access_log = false# rotate access log time base, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 00:00
# since V4.00
access_log_rotate_time = 00:00# if compress the old access log by gzip
# default value is false
# since V6.04
compress_old_access_log = false# compress the access log days before
# default value is 1
# since V6.04
compress_access_log_days_before = 7# if rotate the error log every day
# default value is false
# since V4.02
rotate_error_log = false# rotate error log time base, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 00:00
# since V4.02
error_log_rotate_time = 00:00# if compress the old error log by gzip
# default value is false
# since V6.04
compress_old_error_log = false# compress the error log days before
# default value is 1
# since V6.04
compress_error_log_days_before = 7# rotate access log when the log file exceeds this size
# 0 means never rotates log file by log file size
# default value is 0
# since V4.02
rotate_access_log_size = 0# rotate error log when the log file exceeds this size
# 0 means never rotates log file by log file size
# default value is 0
# since V4.02
rotate_error_log_size = 0# keep days of the log files
# 0 means do not delete old log files
# default value is 0
log_file_keep_days = 0# if skip the invalid record when sync file
# default value is false
# since V4.02
file_sync_skip_invalid_record = false# if use connection pool
# default value is false
# since V4.05
use_connection_pool = true# connections whose the idle time exceeds this time will be closed
# unit: second
# default value is 3600
# since V4.05
connection_pool_max_idle_time = 3600# if compress the binlog files by gzip
# default value is false
# since V6.01
compress_binlog = true# try to compress binlog time, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 01:30
# since V6.01
compress_binlog_time = 01:30# if check the mark of store path to prevent confusion
# recommend to set this parameter to true
# if two storage servers (instances) MUST use a same store path for
# some specific purposes, you should set this parameter to false
# default value is true
# since V6.03
check_store_path_mark = true# use the ip address of this storage server if domain_name is empty,
# else this domain name will ocur in the url redirected by the tracker server
http.domain_name =# the port of the web server on this storage server
http.server_port = 80

1.8.3配置nginx

/etc/fdfs/mod_fastdfs.conf
修改I:
tracker_server=192.168.52.1:22122 #tracker服务器IP和端口
url_have_group_name=true
store_path0=/home/dfs/resources

# connect timeout in seconds
# default value is 30s
connect_timeout=2# network recv and send timeout in seconds
# default value is 30s
network_timeout=30# the base path to store log files
base_path=/tmp# if load FastDFS parameters from tracker server
# since V1.12
# default value is false
load_fdfs_parameters_from_tracker=true# storage sync file max delay seconds
# same as tracker.conf
# valid only when load_fdfs_parameters_from_tracker is false
# since V1.12
# default value is 86400 seconds (one day)
storage_sync_file_max_delay = 86400# if use storage ID instead of IP address
# same as tracker.conf
# valid only when load_fdfs_parameters_from_tracker is false
# default value is false
# since V1.13
use_storage_id = false# specify storage ids filename, can use relative or absolute path
# same as tracker.conf
# valid only when load_fdfs_parameters_from_tracker is false
# since V1.13
storage_ids_filename = storage_ids.conf# FastDFS tracker_server can ocur more than once, and tracker_server format is
#  "host:port", host can be hostname or ip address
# valid only when load_fdfs_parameters_from_tracker is true
tracker_server=192.168.177.128:22122# the port of the local storage server
# the default value is 23000
storage_server_port=23000# the group name of the local storage server
group_name=group1# if the url / uri including the group name
# set to false when uri like /M00/00/00/xxx
# set to true when uri like ${group_name}/M00/00/00/xxx, such as group1/M00/xxx
# default value is false
url_have_group_name = true# path(disk or mount point) count, default value is 1
# must same as storage.conf
store_path_count=1# store_path#, based 0, if store_path0 not exists, it's value is base_path
# the paths must be exist
# must same as storage.conf
store_path0=/home/dfs/resources
#store_path1=/home/yuqing/fastdfs1# standard log level as syslog, case insensitive, value list:
### emerg for emergency
### alert
### crit for critical
### error
### warn for warning
### notice
### info
### debug
log_level=info# set the log filename, such as /usr/local/apache2/logs/mod_fastdfs.log
# empty for output to stderr (apache and nginx error_log file)
log_filename=# response mode when the file not exist in the local file system
## proxy: get the content from other storage server, then send to client
## redirect: redirect to the original storage server (HTTP Header is Location)
response_mode=proxy# the NIC alias prefix, such as eth in Linux, you can see it by ifconfig -a
# multi aliases split by comma. empty value means auto set by OS type
# this paramter used to get all ip address of the local host
# default values is empty
if_alias_prefix=# use "#include" directive to include HTTP config file
# NOTE: #include is an include directive, do NOT remove the # before include
#include http.conf# if support flv
# default value is false
# since v1.15
flv_support = true# flv file extension name
# default value is flv
# since v1.15
flv_extension = flv# set the group count
# set to none zero to support multi-group on this storage server
# set to 0  for single group only
# groups settings section as [group1], [group2], ..., [groupN]
# default value is 0
# since v1.14
group_count = 0# group settings for group #1
# since v1.14
# when support multi-group on this storage server, uncomment following section
#[group1]
#group_name=group1
#storage_server_port=23000
#store_path_count=2
#store_path0=/home/yuqing/fastdfs
#store_path1=/home/yuqing/fastdfs1# group settings for group #2
# since v1.14
# when support multi-group, uncomment following section as neccessary
#[group2]
#group_name=group2
#storage_server_port=23000
#store_path_count=1
#store_path0=/home/yuqing/fastdfs

/usr/local/nginx/conf/nginx.conf
#添加如下配置
server {
listen 80; ## 该端口为storage.conf中的http.server_port相同
server_name localhost;
location ~/group[0-9]/ {
ngx_fastdfs_module;
}
error_page 500 502 503 504 /50x.html;
location = /50x.html {
root html;
}
}


#user  nobody;
worker_processes  1;#error_log  logs/error.log;
#error_log  logs/error.log  notice;
#error_log  logs/error.log  info;#pid        logs/nginx.pid;events {worker_connections  1024;
}http {include       mime.types;default_type  application/octet-stream;#log_format  main  '$remote_addr - $remote_user [$time_local] "$request" '#                  '$status $body_bytes_sent "$http_referer" '#                  '"$http_user_agent" "$http_x_forwarded_for"';#access_log  logs/access.log  main;sendfile        on;#tcp_nopush     on;#keepalive_timeout  0;keepalive_timeout  65;#gzip  on;server {listen       80;server_name  localhost;#charset koi8-r;#access_log  logs/host.access.log  main;location / {root   html;index  index.html index.htm;}location ~/group[0-9]/ {ngx_fastdfs_module;}#error_page  404              /404.html;# redirect server error pages to the static page /50x.html#error_page   500 502 503 504  /50x.html;location = /50x.html {root   html;}# proxy the PHP scripts to Apache listening on 127.0.0.1:80##location ~ \.php$ {#    proxy_pass   http://127.0.0.1;#}# pass the PHP scripts to FastCGI server listening on 127.0.0.1:9000##location ~ \.php$ {#    root           html;#    fastcgi_pass   127.0.0.1:9000;#    fastcgi_index  index.php;#    fastcgi_param  SCRIPT_FILENAME  /scripts$fastcgi_script_name;#    include        fastcgi_params;#}# deny access to .htaccess files, if Apache's document root# concurs with nginx's one##location ~ /\.ht {#    deny  all;#}}# another virtual host using mix of IP-, name-, and port-based configuration##server {#    listen       8000;#    listen       somename:8080;#    server_name  somename  alias  another.alias;#    location / {#        root   html;#        index  index.html index.htm;#    }#}# HTTPS server##server {#    listen       443 ssl;#    server_name  localhost;#    ssl_certificate      cert.pem;#    ssl_certificate_key  cert.key;#    ssl_session_cache    shared:SSL:1m;#    ssl_session_timeout  5m;#    ssl_ciphers  HIGH:!aNULL:!MD5;#    ssl_prefer_server_ciphers  on;#    location / {#        root   html;#        index  index.html index.htm;#    }#}}

2启动

2.1关闭防火墙

#不关闭防火墙的话无法使用
systemctl stop firewalld.service #关闭
systemctl restart firewalld.service #重启

2.2启动tracker

/etc/init.d/fdfs_trackerd start #启动tracker服务
/etc/init.d/fdfs_trackerd restart #重启动tracker服务
/etc/init.d/fdfs_trackerd stop #停止tracker服务
chkconfig fdfs_trackerd on #自启动tracker服务

2.3启动storage

/etc/init.d/fdfs_storaged start #启动storage服务
/etc/init.d/fdfs_storaged restart #重动storage服务
/etc/init.d/fdfs_storaged stop #停止动storage服务
chkconfig fdfs_storaged on #自启动storage服务

2.4启动nginx

/usr/local/nginx/sbin/nginx #启动nginx
/usr/local/nginx/sbin/nginx -s reload #重启nginx
/usr/local/nginx/sbin/nginx -s stop #停止nginx

2.4检测

/usr/bin/fdfs_monitor /etc/fdfs/storage.conf
# 会显示会有几台服务器 有3台就会 显示 Storage 1-Storage 3的详细信息

配置文件说明

tracker_server #有几台服务器写几个
group_name #地址的名称的命名
bind_addr #服务器ip绑定
store_path_count #store_path(数字)有几个写几个
store_path(数字) #设置几个储存地址写几个 从0开始

3.java fastDFS cilent

下载源码:https://github.com/happyfish100/fastdfs-client-java
将源码打成jar包
导入jar包

<dependency><groupId>org.csource</groupId><artifactId>fastdfs-client-java</artifactId><exclusions><exclusion><groupId>org.slf4j</groupId><artifactId>slf4j-log4j12</artifactId></exclusion></exclusions></dependency>

工具类

package com.wal.hrm.utils;import org.csource.common.NameValuePair;
import org.csource.fastdfs.*;public class FastDfsApiOpr {public static String CONF_FILENAME  = FastDfsApiOpr.class.getClassLoader().getResource("fast_client.conf").getFile();/*** 上传文件* @param file* @param extName* @return*/public static  String upload(byte[] file,String extName) {try {ClientGlobal.init(CONF_FILENAME);TrackerClient tracker = new TrackerClient();TrackerServer trackerServer = tracker.getTrackerServer();StorageServer storageServer = null;StorageClient storageClient = new StorageClient(trackerServer, storageServer);/* NameValuePair nvp [] = new NameValuePair[]{new NameValuePair("age", "18"),new NameValuePair("sex", "male")};*/String fileIds[] = storageClient.upload_file(file,extName,null);System.out.println(fileIds.length);System.out.println("组名:" + fileIds[0]);System.out.println("路径: " + fileIds[1]);return  "/"+fileIds[0]+"/"+fileIds[1];} catch (Exception e) {e.printStackTrace();return  null;}}/*** 上传文件* @param extName* @return*/public static  String upload(String path,String extName) {try { ClientGlobal.init(CONF_FILENAME);TrackerClient tracker = new TrackerClient(); TrackerServer trackerServer = tracker.getTrackerServer();StorageServer storageServer = null;StorageClient storageClient = new StorageClient(trackerServer, storageServer);String fileIds[] = storageClient.upload_file(path, extName,null);System.out.println(fileIds.length); System.out.println("组名:" + fileIds[0]); System.out.println("路径: " + fileIds[1]);return  "/"+fileIds[0]+"/"+fileIds[1];} catch (Exception e) {e.printStackTrace();return  null;}}/*** 下载文件* @param groupName* @param fileName* @return*/public static byte[] download(String groupName,String fileName) {try {ClientGlobal.init(CONF_FILENAME);TrackerClient tracker = new TrackerClient(); TrackerServer trackerServer = tracker.getTrackerServer();StorageServer storageServer = null;StorageClient storageClient = new StorageClient(trackerServer, storageServer); byte[] b = storageClient.download_file(groupName, fileName);return  b;} catch (Exception e) {e.printStackTrace();return  null;} }/*** 删除文件* @param groupName* @param fileName*/public static void delete(String groupName,String fileName){try { ClientGlobal.init(CONF_FILENAME);TrackerClient tracker = new TrackerClient(); TrackerServer trackerServer = tracker.getTrackerServer();StorageServer storageServer = null;StorageClient storageClient = new StorageClient(trackerServer, storageServer); int i = storageClient.delete_file(groupName,fileName);System.out.println( i==0 ? "删除成功" : "删除失败:"+i);} catch (Exception e) {e.printStackTrace();throw  new RuntimeException("删除异常,"+e.getMessage());} }
}

fast_client.conf配置

tracker_server=192.168.177.128:22122

fastfds文件服务器搭建相关推荐

  1. 文件服务器搭建 xp,xp文件服务器搭建

    xp文件服务器搭建 内容精选 换一换 登录Windows操作系统的弹性云服务器时,需使用密码方式登录.因此,用户需先根据创建弹性云服务器时使用的密钥文件,获取该弹性云服务器初始安装时系统生成的管理员密 ...

  2. 文件服务器 说明,文件服务器搭建说明.pdf

    文件共享服务器搭建 责任人 姜 源 起讫时间 2010 11 15 至 2010 12 19 编号 JY001 0 目录目录 一 确定实现方案 1 二 方案实现方法 4 三 参考资料 12 1 文件服 ...

  3. Nginx 静态文件服务器搭建及autoindex模块解析

    ngx_http_autoindex_module ngx_http_autoindex_module模块处理以斜杠字符('/')结尾的请求,并生成目录列表. 当ngx_http_index_modu ...

  4. 企业文件服务器搭建案例

    6.9 章节案例 在本章案例讲解中通过4个模拟项目对Samba的功能进行演示,其中第一个案例模拟一个企业中服务器全部基于RHEL5.x,客户端在访问Samba服务器时,用户认证及所有用户行为控制全部通 ...

  5. windows文件服务器 文件方案,windowsserver2008文件服务器搭建2种方案.docx

    文件服务器搭建的两种方案范光华制作 文件服务器搭建的两种方案范光华制作 文件服务器搭建的两种方案 搭建目的: 1.实现企业文件共享与传输,提高工作效率: 2.提高企业访问文件的安全性. 搭建环境: 1 ...

  6. 企业架设文件服务器,企业文件服务器搭建案例

    <Linux服务器配置全程实录>第6章使用Samba配置文件服务器.Samba作为类UNIX系统和Windows的通信的桥梁,目前应用非常广泛,本章中在RHEL下通过Samba配置文件服务 ...

  7. 阿里云服务器ECS 第一篇:FTP文件服务器搭建

    < 第一篇:FTP文件服务器搭建 > 目前打算将项目部署到阿里云服务器上,并配置系统运行环境,搭建FTP服务器是系统部署的第一步,为了后期文件传输方便快捷,所以从这里开始!!! 关于FTP ...

  8. 邮件服务器 文件服务器,搭建邮件、终端和文件服务器应用方案_服务器_服务器x86服务器-中关村在线...

    邮件服务器方案 通常来说,初创企业邮件服务器对处理器性能要求并不高,但是由于要支持一定数量的并发连接,对于网络子系统和内存有一定的要求.邮件服务器软件对于内存需求也较高,同时需要较大的存储空间用来邮件 ...

  9. 【tool】企业级开源分布式文件服务器搭建(FastDFS)

    项目介绍 FastDFS介绍 FastDFS是一个以C语言开发的开源轻量级分布式文件系统,由阿里巴巴开发并开源.它对文件进行管理,功能包括:文件存储.文件同步.文件访问(上传.下载)等.解决了大容量存 ...

  10. 文件服务器搭建_小型企业文件存储服务器的搭建四部曲

    随着科技的发展,越来越多的企业都选用了服务器,但是很多企业为了节省成本,购买存储服务器并不是最好的选择,所以这就使得很多企业自己私底下就搭建了企业文件存储服务器.今天小编就给大家带来这四个步骤: 使用 ...

最新文章

  1. 关于客户需求分析的讨论
  2. [唐诗]古风(其二十四)-李白
  3. IBASE component deletion
  4. matlab 泡泡图,使用matlab绘制2维、3维气泡图
  5. java jexl 工具类_jexl表达式解析、计算工具类.md
  6. iptables基本概念详解
  7. Andrew ng清华报告听后感
  8. ValueError: output parameter for reduction operation logical_and has too many dimensions ?
  9. java获得时间差_Java获取时间差实例代码详解
  10. Garbled Circuits介绍 - 2 基础知识
  11. LDA-线性判别分析
  12. Word中部分内容出现乱码、不显示页码和目录
  13. 简单介绍公有云、私有云、混合云
  14. win10 win11睡眠后系统无法唤醒解决办法
  15. prometheus监控-1
  16. 微信公告号 图灵机器人实现智能回复
  17. DirectShow使用大全
  18. 广义相对论基础【3】平移联络
  19. 华硕笔记本触控板有linux驱动吗,华硕笔记本怎么下载触控板驱动
  20. 面试经验 - 如何介绍自己的技能水平?

热门文章

  1. 手机号检测性别原理分析 微信男女 抖音ID检测原理
  2. 斯坦福发布AI年度报告,中国AI论文数全球第二,第一也不是美国【附下载】
  3. 用java写图形验证码,超级简单
  4. 树莓派4B全40管脚对应功能示意图
  5. java插桩-javaassist
  6. NB-IoT未来发展,主要是靠什么驱动的?
  7. 启动项目时出现java.io.EOFException异常
  8. android镜子app,Android镜子应用 一面可编程的镜子
  9. 需求管理搞不定?这4招帮你解决项目需求管理
  10. linux 繁体转简体,Linux下在程序中如何进行繁体中文和简体中文的转换