Skip to content

Images 0.9.x

Images
  • Hallo, ich habe noch intermediate Release von den 0.8.x
    Kann ich von denen auf das aktuelle upgraden?
    Gruß,
    Christoph

  • Hallo @cnaed ,

    Release Versionen aktualisiert er ja sowieso, auf eine Pre-Release Version zu aktualisieren bedarf etwas Handarbeit.

    nano /etc/apt/sources.list.d/ayufan-rock64.list
    

    Inhalt der Datei

    deb http://deb.ayufan.eu/orgs/ayufan-rock64/releases /
    
    # uncomment to use pre-release kernels and compatibility packages
    # deb http://deb.ayufan.eu/orgs/ayufan-rock64/pre-releases /
    

    Sollte selbsterklärend sein 😉

    Bei produktiven Systemen, vorher zweimal nachdenken, ob es sinnvoll ist das Update/Upgrade zu machen. Es kommt zwar aktuell selten vor, aber es gibt immer noch Probleme. z.B.: meine aktuelle Testinstallation mit SPI und USB3-Boot knallt ständig mit haufenweisen ext4 Fehler usw. Nicht so gut!

  • 0.9.4: gitlab-ci-linux-build-146 released

    • 0.9.4: Update kernel branch to 4.4.189,

  • Aktuell sind wir im Moment bei 0.9.8: gitlab-ci-linux-build-151

    • 0.9.8: Fix BT audio (run on 1.5M),
    • 0.9.8: Use DMA on UART/SPI,
    • 0.9.8: Lower temperature thresholds for Pinebook Pro,
    • 0.9.8: Set default audio device for all devices,
    • 0.9.8: Fix OMV4 build,
    • 0.9.8: Automatically enable h264ify,
    • 0.9.7: Include a list of used packages,
    • 0.9.7: Fix regression on compositing performance,
    • 0.9.7: Force password to change on first login,
    • 0.9.6: Support Pinebook Pro v2.1,
    • 0.9.6: Optimise suspend on Pinebook Pro,
    • 0.9.6: Optimise Mate styling,
    • 0.9.6: Optimise Touchpad settings for Pinebook Pro,
    • 0.9.6: Support 2/1.5GHz OPP on Pinebook Pro,
    • 0.9.5: Build Debian/Stretch/OMV4,
    • 0.9.5: Improve install_container_linux.sh,

    Quelle: https://github.com/ayufan-rock64/linux-build/releases/tag/0.9.8

  • 0.9.9: gitlab-ci-linux-build-152 released

    • 0.9.9: Fix Firefox video playback,
    • 0.9.9: Remove libmali-rk-dev from default install,
    • 0.9.9: Align standby/work leds across all boards,
    • 0.9.9: Disable Debug UART on Pinebook Pro, as it causes stability issues,
    • 0.9.9: Fix Pinebook Pro SD card stability,
    • 0.9.9: Enable PCIE NVME support for Pinebook Pro,
    • 0.9.8: Fix BT audio (run on 1.5M),
  • 0.9.10: gitlab-ci-linux-build-154 released

    • 0.9.10: Fix support for power/standby LEDs for all boards,
    • 0.9.10: Fix rock64 gpu acceleration regression introduced in 0.9.9,
    • 0.9.10: Replace lxdm to use lightdm, as this allows password change on login,
    • 0.9.10: Remove gnome-screensaver to fix double lock screen,
  • 0.9.11: gitlab-ci-linux-build-155 released

    • 0.9.11: Install unity-greeter,
  • 0.9.12: gitlab-ci-linux-build-156 released

    • 0.9.12: Fix LXDE for Rock64,
  • 0.9.13: gitlab-ci-linux-build-157 released

    • 0.9.13: Bump sound volume for Pinebook Pro,
    • 0.9.13: Fix Firefox video playback,
  • 0.9.14: gitlab-ci-linux-build-159 released

    • 0.9.14: Bump kernel to 4.4.190,
    • 0.9.14: Fix Firefox video playback,
  • 0.9.16: gitlab-ci-linux-build-163 released

    0.9.x

    • 0.9.16: Bump kernel to 4.4.197,
    • 0.9.15: Bump kernel to 4.4.193,
    • 0.9.14: Bump kernel to 4.4.190,
    • 0.9.14: Fix Firefox video playback,
    • 0.9.13: Bump sound volume for Pinebook Pro,
    • 0.9.12: Fix LXDE for Rock64,
    • 0.9.10: Fix support for power/standby LEDs for all boards,

  • [V] ROCKPro64 incl. PCIe SATA-Karte

    Verschoben Archiv
    2
    +1
    0 Stimmen
    2 Beiträge
    261 Aufrufe
    FrankMF
    Verkauft!
  • ROCKPro64 - RTL8111/8168/8411 Netzwerkkarte

    Hardware
    4
    +0
    0 Stimmen
    4 Beiträge
    405 Aufrufe
    K
    na denn, tippe ich mal so auf default konfiguriert per dhcp
  • ROCKPro64 - Release 0.10.6 Was geht?

    Angeheftet ROCKPro64
    1
    0 Stimmen
    1 Beiträge
    407 Aufrufe
    Niemand hat geantwortet
  • Serielle Konsole UART2 (2)

    Angeheftet Hardware
    1
    +7
    0 Stimmen
    1 Beiträge
    229 Aufrufe
    Niemand hat geantwortet
  • Armbian 5.4.0-rc1

    Armbian
    5
    +3
    0 Stimmen
    5 Beiträge
    386 Aufrufe
    FrankMF
    Gut, ich bin nicht der einzige, der ständig damit Probleme hat. @tkaiser auch [image: 1578061215343-1036201d-a4b2-47be-a618-36003c07e0ce-grafik.png]
  • ROCKPro64 Übersicht - was geht? **veraltet**

    Angeheftet Verschoben Archiv
    5
    0 Stimmen
    5 Beiträge
    2k Aufrufe
    FrankMF
    Ich sehe gerade, das könnte hier auch mal neu gemacht werden.
  • Benchmark Mainline 4.17.0-rc6

    Verschoben Archiv
    4
    0 Stimmen
    4 Beiträge
    1k Aufrufe
    FrankMF
    iozone 5GT/s x2 rock64@rockpro64:/mnt$ sudo iozone -e -I -a -s 100M -r 4k -r 16k -r 512k -r 1024k -r 16384k -i 0 -i 1 -i 2 Iozone: Performance Test of File I/O Version $Revision: 3.429 $ Compiled for 64 bit mode. Build: linux Contributors:William Norcott, Don Capps, Isom Crawford, Kirby Collins Al Slater, Scott Rhine, Mike Wisner, Ken Goss Steve Landherr, Brad Smith, Mark Kelly, Dr. Alain CYR, Randy Dunlap, Mark Montague, Dan Million, Gavin Brebner, Jean-Marc Zucconi, Jeff Blomberg, Benny Halevy, Dave Boone, Erik Habbinga, Kris Strecker, Walter Wong, Joshua Root, Fabrice Bacchella, Zhenghua Xue, Qin Li, Darren Sawyer, Vangel Bojaxhi, Ben England, Vikentsi Lapa. Run began: Sat Jun 16 06:34:43 2018 Include fsync in write timing O_DIRECT feature enabled Auto Mode File size set to 102400 kB Record Size 4 kB Record Size 16 kB Record Size 512 kB Record Size 1024 kB Record Size 16384 kB Command line used: iozone -e -I -a -s 100M -r 4k -r 16k -r 512k -r 1024k -r 16384k -i 0 -i 1 -i 2 Output is in kBytes/sec Time Resolution = 0.000001 seconds. Processor cache size set to 1024 kBytes. Processor cache line size set to 32 bytes. File stride size set to 17 * record size. random random bkwd record stride kB reclen write rewrite read reread read write read rewrite read fwrite frewrite fread freread 102400 4 48672 104754 115838 116803 47894 103606 102400 16 168084 276437 292660 295458 162550 273703 102400 512 566572 597648 580005 589209 534508 597007 102400 1024 585621 624443 590545 599177 569452 630098 102400 16384 504871 754710 765558 780592 777696 753426 iozone test complete. 2,5GT/s x2 rock64@rockpro64:/mnt$ sudo iozone -e -I -a -s 100M -r 4k -r 16k -r 512k -r 1024k -r 16384k -i 0 -i 1 -i 2 Iozone: Performance Test of File I/O Version $Revision: 3.429 $ Compiled for 64 bit mode. Build: linux Contributors:William Norcott, Don Capps, Isom Crawford, Kirby Collins Al Slater, Scott Rhine, Mike Wisner, Ken Goss Steve Landherr, Brad Smith, Mark Kelly, Dr. Alain CYR, Randy Dunlap, Mark Montague, Dan Million, Gavin Brebner, Jean-Marc Zucconi, Jeff Blomberg, Benny Halevy, Dave Boone, Erik Habbinga, Kris Strecker, Walter Wong, Joshua Root, Fabrice Bacchella, Zhenghua Xue, Qin Li, Darren Sawyer, Vangel Bojaxhi, Ben England, Vikentsi Lapa. Run began: Sun Jun 17 06:54:02 2018 Include fsync in write timing O_DIRECT feature enabled Auto Mode File size set to 102400 kB Record Size 4 kB Record Size 16 kB Record Size 512 kB Record Size 1024 kB Record Size 16384 kB Command line used: iozone -e -I -a -s 100M -r 4k -r 16k -r 512k -r 1024k -r 16384k -i 0 -i 1 -i 2 Output is in kBytes/sec Time Resolution = 0.000001 seconds. Processor cache size set to 1024 kBytes. Processor cache line size set to 32 bytes. File stride size set to 17 * record size. random random bkwd record stride kB reclen write rewrite read reread read write read rewrite read fwrite frewrite fread freread 102400 4 49420 91310 102658 103415 47023 90099 102400 16 138141 202088 224648 225918 141642 202457 102400 512 335055 347517 375096 378596 364668 350005 102400 1024 345508 354999 378947 382733 375315 354783 102400 16384 306262 383155 424403 429423 428670 377476 iozone test complete.
  • stretch-minimal-rockpro64

    Verschoben Linux
    3
    0 Stimmen
    3 Beiträge
    1k Aufrufe
    FrankMF
    Mal ein Test was der Speicher so kann. rock64@rockpro64:~/tinymembench$ ./tinymembench tinymembench v0.4.9 (simple benchmark for memory throughput and latency) ========================================================================== == Memory bandwidth tests == == == == Note 1: 1MB = 1000000 bytes == == Note 2: Results for 'copy' tests show how many bytes can be == == copied per second (adding together read and writen == == bytes would have provided twice higher numbers) == == Note 3: 2-pass copy means that we are using a small temporary buffer == == to first fetch data into it, and only then write it to the == == destination (source -> L1 cache, L1 cache -> destination) == == Note 4: If sample standard deviation exceeds 0.1%, it is shown in == == brackets == ========================================================================== C copy backwards : 2812.7 MB/s C copy backwards (32 byte blocks) : 2811.9 MB/s C copy backwards (64 byte blocks) : 2632.8 MB/s C copy : 2667.2 MB/s C copy prefetched (32 bytes step) : 2633.5 MB/s C copy prefetched (64 bytes step) : 2640.8 MB/s C 2-pass copy : 2509.8 MB/s C 2-pass copy prefetched (32 bytes step) : 2431.6 MB/s C 2-pass copy prefetched (64 bytes step) : 2424.1 MB/s C fill : 4887.7 MB/s (0.5%) C fill (shuffle within 16 byte blocks) : 4883.0 MB/s C fill (shuffle within 32 byte blocks) : 4889.3 MB/s C fill (shuffle within 64 byte blocks) : 4889.2 MB/s --- standard memcpy : 2807.3 MB/s standard memset : 4890.4 MB/s (0.3%) --- NEON LDP/STP copy : 2803.7 MB/s NEON LDP/STP copy pldl2strm (32 bytes step) : 2802.1 MB/s NEON LDP/STP copy pldl2strm (64 bytes step) : 2800.7 MB/s NEON LDP/STP copy pldl1keep (32 bytes step) : 2745.5 MB/s NEON LDP/STP copy pldl1keep (64 bytes step) : 2745.8 MB/s NEON LD1/ST1 copy : 2801.9 MB/s NEON STP fill : 4888.9 MB/s (0.3%) NEON STNP fill : 4850.1 MB/s ARM LDP/STP copy : 2803.8 MB/s ARM STP fill : 4893.0 MB/s (0.5%) ARM STNP fill : 4851.7 MB/s ========================================================================== == Framebuffer read tests. == == == == Many ARM devices use a part of the system memory as the framebuffer, == == typically mapped as uncached but with write-combining enabled. == == Writes to such framebuffers are quite fast, but reads are much == == slower and very sensitive to the alignment and the selection of == == CPU instructions which are used for accessing memory. == == == == Many x86 systems allocate the framebuffer in the GPU memory, == == accessible for the CPU via a relatively slow PCI-E bus. Moreover, == == PCI-E is asymmetric and handles reads a lot worse than writes. == == == == If uncached framebuffer reads are reasonably fast (at least 100 MB/s == == or preferably >300 MB/s), then using the shadow framebuffer layer == == is not necessary in Xorg DDX drivers, resulting in a nice overall == == performance improvement. For example, the xf86-video-fbturbo DDX == == uses this trick. == ========================================================================== NEON LDP/STP copy (from framebuffer) : 602.5 MB/s NEON LDP/STP 2-pass copy (from framebuffer) : 551.6 MB/s NEON LD1/ST1 copy (from framebuffer) : 667.1 MB/s NEON LD1/ST1 2-pass copy (from framebuffer) : 605.6 MB/s ARM LDP/STP copy (from framebuffer) : 445.3 MB/s ARM LDP/STP 2-pass copy (from framebuffer) : 428.8 MB/s ========================================================================== == Memory latency test == == == == Average time is measured for random memory accesses in the buffers == == of different sizes. The larger is the buffer, the more significant == == are relative contributions of TLB, L1/L2 cache misses and SDRAM == == accesses. For extremely large buffer sizes we are expecting to see == == page table walk with several requests to SDRAM for almost every == == memory access (though 64MiB is not nearly large enough to experience == == this effect to its fullest). == == == == Note 1: All the numbers are representing extra time, which needs to == == be added to L1 cache latency. The cycle timings for L1 cache == == latency can be usually found in the processor documentation. == == Note 2: Dual random read means that we are simultaneously performing == == two independent memory accesses at a time. In the case if == == the memory subsystem can't handle multiple outstanding == == requests, dual random read has the same timings as two == == single reads performed one after another. == ========================================================================== block size : single random read / dual random read 1024 : 0.0 ns / 0.0 ns 2048 : 0.0 ns / 0.0 ns 4096 : 0.0 ns / 0.0 ns 8192 : 0.0 ns / 0.0 ns 16384 : 0.0 ns / 0.0 ns 32768 : 0.0 ns / 0.0 ns 65536 : 4.5 ns / 7.2 ns 131072 : 6.8 ns / 9.7 ns 262144 : 9.8 ns / 12.8 ns 524288 : 11.4 ns / 14.7 ns 1048576 : 16.0 ns / 22.6 ns 2097152 : 114.0 ns / 175.3 ns 4194304 : 161.7 ns / 219.9 ns 8388608 : 190.7 ns / 241.5 ns 16777216 : 205.3 ns / 250.5 ns 33554432 : 212.9 ns / 255.5 ns 67108864 : 222.3 ns / 271.1 ns