Ihor Hosting logo

Добрый день!

Такой сервер я уже тестировал 2 года назад и теперь вот решил попробовать еще разок. В целом все без изменений кроме того что скорость чтения и записи на диск стала сильно выше.

Есть ощущение что там SSD =)

Опубликую только результаты связанные со скоростью записи на диск, остальное в пределах погрешности.

Сведения о VM

PT Summary

pt-summary
-----------
# Percona Toolkit System Summary Report ######################
        Date | 2020-09-04 10:40:50 UTC (local TZ: MSK +0300)
    Hostname | vds1968051.my-ihor.ru
      Uptime | 10 min,  1 user,  load average: 0.04, 0.13, 0.08
      System | Red Hat; KVM; vRHEL 7.6.0 PC (i440FX + PIIX, 1996) (Other)
 Service Tag | Not Specified
    Platform | Linux
     Release | CentOS Linux release 7.8.2003 (Core)
      Kernel | 3.10.0-1127.19.1.el7.x86_64
Architecture | CPU = 64-bit, OS = 64-bit
   Threading | NPTL 2.17
     SELinux | Disabled
 Virtualized | VMWare
# Processor ##################################################
  Processors | physical = 1, cores = 1, virtual = 1, hyperthreading = no
      Speeds | 1x2799.998
      Models | 1xQEMU Virtual CPU version 2.5+
      Caches | 1x16384 KB
# Memory #####################################################
       Total | 486.9M
        Free | 69.2M
        Used | physical = 95.6M, swap allocated = 0.0, swap used = 0.0, virtual = 95.6M
      Shared | 4.4M
     Buffers | 322.0M
      Caches | 374.1M
       Dirty | 92 kB
     UsedRSS | 106.1M
  Swappiness | 30
 DirtyPolicy | 30, 10
 DirtyStatus | 0, 0
  Locator   Size     Speed             Form Factor   Type          Type Detail
  ========= ======== ================= ============= ============= ===========
  DIMM 0    512 MB   Unknown           DIMM          RAM           Other
# Mounted Filesystems ########################################
  Filesystem  Size Used Type     Opts                                            Mountpoint
  devtmpfs    233M   0% devtmpfs rw,nosuid,size=237644k,nr_inodes=59411,mode=755 /dev
  /dev/vda1   488M  23% ext4     rw,relatime,data=ordered                        /boot
  /dev/vda2   9.5G  16% xfs      rw,relatime,attr2,inode64,noquota               /
  tmpfs       244M   0% tmpfs    rw,nosuid,nodev                                 /dev/shm
  tmpfs       244M   0% tmpfs    rw,nosuid,nodev,mode=755                        /dev/shm
  tmpfs       244M   0% tmpfs    rw,nosuid,nodev,relatime,size=49864k,mode=700   /dev/shm
  tmpfs       244M   0% tmpfs    ro,nosuid,nodev,noexec,mode=755                 /dev/shm
  tmpfs       244M   0% tmpfs    rw,nosuid,nodev                                 /sys/fs/cgroup
  tmpfs       244M   0% tmpfs    rw,nosuid,nodev,mode=755                        /sys/fs/cgroup
  tmpfs       244M   0% tmpfs    rw,nosuid,nodev,relatime,size=49864k,mode=700   /sys/fs/cgroup
  tmpfs       244M   0% tmpfs    ro,nosuid,nodev,noexec,mode=755                 /sys/fs/cgroup
  tmpfs       244M   2% tmpfs    rw,nosuid,nodev                                 /run
  tmpfs       244M   2% tmpfs    rw,nosuid,nodev,mode=755                        /run
  tmpfs       244M   2% tmpfs    rw,nosuid,nodev,relatime,size=49864k,mode=700   /run
  tmpfs       244M   2% tmpfs    ro,nosuid,nodev,noexec,mode=755                 /run
  tmpfs        49M   0% tmpfs    rw,nosuid,nodev                                 /run/user/0
  tmpfs        49M   0% tmpfs    rw,nosuid,nodev,mode=755                        /run/user/0
  tmpfs        49M   0% tmpfs    rw,nosuid,nodev,relatime,size=49864k,mode=700   /run/user/0
  tmpfs        49M   0% tmpfs    ro,nosuid,nodev,noexec,mode=755                 /run/user/0
# Disk Schedulers And Queue Size #############################
         vda | [mq-deadline] 256
# Disk Partioning ############################################
Device       Type      Start        End               Size
============ ==== ========== ========== ==================
/dev/vda     Disk                              10737418240
/dev/vda1    Part       2048    1050623          536870400
/dev/vda2    Part    1050624   20971519        10199498240
# Kernel Inode State #########################################
dentry-state | 30377	19674	45	0	9185	0
     file-nr | 928	0	46628
    inode-nr | 20189	102
# LVM Volumes ################################################
Unable to collect information
# LVM Volume Groups ##########################################
Unable to collect information
# RAID Controller ############################################
  Controller | No RAID controller detected
# Network Config #############################################
  Controller | Red Hat, Inc. Virtio network device
 FIN Timeout | 60
  Port Range | 60999
# Interface Statistics #######################################
  interface  rx_bytes rx_packets  rx_errors   tx_bytes tx_packets  tx_errors
  ========= ========= ========== ========== ========== ========== ==========
  lo                0          0          0          0          0          0
  eth0       60000000      40000          0    2250000      22500          0
# Network Devices ############################################
  Device    Speed     Duplex
  ========= ========= =========
  eth0
# Top Processes ##############################################
  PID USER      PR  NI    VIRT    RES    SHR S %CPU %MEM     TIME+ COMMAND
    1 root      20   0   43580   3776   2472 S  0.0  0.8   0:01.14 systemd
    2 root      20   0       0      0      0 S  0.0  0.0   0:00.00 kthreadd
    4 root       0 -20       0      0      0 S  0.0  0.0   0:00.00 kworker/0:0H
    5 root      20   0       0      0      0 S  0.0  0.0   0:00.09 kworker/u2:0
    6 root      20   0       0      0      0 S  0.0  0.0   0:00.05 ksoftirqd/0
    7 root      rt   0       0      0      0 S  0.0  0.0   0:00.00 migration/0
    8 root      20   0       0      0      0 S  0.0  0.0   0:00.00 rcu_bh
    9 root      20   0       0      0      0 R  0.0  0.0   0:00.15 rcu_sched
   10 root       0 -20       0      0      0 S  0.0  0.0   0:00.00 lru-add-dra+
# Notable Processes ##########################################
  PID    OOM    COMMAND
  810    -17    sshd
# Memory mamagement ##########################################
Transparent huge pages are currently disabled on the system.
# The End ####################################################

Тесты

DD

dd if=/dev/zero of=test bs=64k count=16k conv=fdatasync 2>&1
rm -f test
-----------
16384+0 records in
16384+0 records out
1073741824 bytes (1.1 GB) copied, 2.1473 s, 500 MB/s
16384+0 records in
16384+0 records out
1073741824 bytes (1.1 GB) copied, 1.64194 s, 654 MB/s
16384+0 records in
16384+0 records out
1073741824 bytes (1.1 GB) copied, 1.71044 s, 628 MB/s

Это в 10 раз быстрее того что было до этого.

Sysbench тест диска

sysbench fileio --file-total-size=1G prepare
-----------
sysbench 1.0.17 (using system LuaJIT 2.0.4)

128 files, 8192Kb each, 1024Mb total
Creating files for the test...
Extra file open flags: (none)
Creating file test_file.0
...
Creating file test_file.127
1073741824 bytes written in 3.61 seconds (283.34 MiB/sec).
sysbench fileio --file-total-size=1G --file-test-mode=rndrw --time=300 --max-requests=0 run
-----------
sysbench 1.0.17 (using system LuaJIT 2.0.4)

Running the test with following options:
Number of threads: 1
Initializing random number generator from current time


Extra file open flags: (none)
128 files, 8MiB each
1GiB total file size
Block size 16KiB
Number of IO requests: 0
Read/Write ratio for combined random IO test: 1.50
Periodic FSYNC enabled, calling fsync() each 100 requests.
Calling fsync() at the end of test, Enabled.
Using synchronous I/O mode
Doing random r/w test
Initializing worker threads...

Threads started!


File operations:
    reads/s:                      1825.15
    writes/s:                     1216.77
    fsyncs/s:                     3893.78

Throughput:
    read, MiB/s:                  28.52
    written, MiB/s:               19.01

General statistics:
    total time:                          300.0213s
    total number of events:              2080799

Latency (ms):
         min:                                    0.00
         avg:                                    0.14
         max:                                  136.85
         95th percentile:                        0.50
         sum:                               296767.58

Threads fairness:
    events (avg/stddev):           2080799.0000/0.00
    execution time (avg/stddev):   296.7676/0.00

Выводы

Взял себе такой сервер чтобы запускать на нем автоматическую сборку веб-приложения. Работает все достаточно шустро для таких скромных показателей. С возросшей производительностью диска это очень хороший вариант за 100 рублей в месяц.