Skip to content

Images 0.7.x

Images
  • Kamil ist wieder am schrauben.

    0.7.x

    • 0.7.0: Introduces heavy refactor splitting all components into separate repos, and separate independent releases (u-boot, kernel, kernel-mainline, compatibility package),
    • 0.7.0: Dry run everything,

    Bin gespannt auf das Ergebnis, im Moment hat das Script mit Fehlern abgebrochen, aber Kamil kennt das bestens 😉

  • Es gibt noch nicht Neues, aber hier kann man schon ein wenig sehen was er macht.

    • 0.7.3: Speed-up build process,
    • 0.7.2: Pin packages,
    • 0.7.1: Use GitLab CI for releasing all images,
    • 0.7.0: Introduces heavy refactor splitting all components into separate repos, and separate independent releases (u-boot, kernel, kernel-mainline, compatibility package),

    LATEST_UBOOT_VERSION=2017.09-rockchip-ayufan-1009-g501b20dc14
    LATEST_KERNEL_VERSION=4.4.132-1042-rockchip-ayufan-g521c26bc8ee0
    LATEST_PACKAGE_VERSION=0.7-5

    Quelle: https://gitlab.com/ayufan-repos/rock64/linux-build/commit/b557556d38760cfd1f7e9a818e7a51c873520161

  • Wenn ich das jetzt richtig beobachtet habe, hat Kamil ganze zwei Tage damit zugebracht, den Build Prozeß umzubauen. Da läuft jetzt wohl alles auf gitlab ab, die Releases findet man aber immer noch hier. Was da wie, womit zusammen arbeitet, kann ich Euch nicht sagen. Bin auf dem Gebiet mit Sicherheit kein Experte. 😉

    Das Gute, es gibt einen neuen Release. 0.7.2

    Das Schlechte, bootet nicht, da scheint was mit dem u-boot nicht zu passen.

    Kamil schrieb

    (23:24:08) ayufan1: I will fix it later
    (23:24:15) ayufan1: I basically changed 80% of the build process 🙂
    (23:24:21) ayufan1: so I expect something not to work 😛

    Dann warten wir mal..

    und DANKE für die Arbeit - Kamil!!

  • Gestern Nacht kam noch 0.7.3

    • 0.7.3: Fix generation of extlinux.conf (linux booting),

    Das Problem des nicht Bootens wurde damit behoben, aber das Image ist unstabil. Kamil weiß das , ich bin sicher das er heute weiter dran rumschraubt.

    Dann mal dran erinnern was im Moment gut funktioniert.

    • 0.6.58: jenkins-linux-build-rock-64-271 installieren
    • und dann auf 4.18.0-rc3-1035-ayufan upgraden

    Die Releases werden übrigens weiterhin auf github veröffentlicht!

  • 0.7.3 released

    • 0.7.3: Fix generation of extlinux.conf (linux booting),
    • 0.7.2: Pin packages,
    • 0.7.2: Improve performance of build process,
    • 0.7.1: Use GitLab CI for releasing all images,
    • 0.7.0: Introduces heavy refactor splitting all components into separate repos, and separate independent releases (u-boot, kernel, kernel-mainline, compatibility package),
    • 0.7.0: Dry run everything,

    Kamil hat seinen kompletten Build-Prozeß umgebaut. Diese 0.7.3 Version bootet (nur ohne PCIe NVMe-Karte). Außerdem habe ich Unstabilitäten über die LAN-Schnittstelle bemerkt (Freeze). Nicht empfehlenswert, im Moment.

  • 0.7.4 released

    • 0.7.4: Fix resize_rootfs.sh script to respect boot flags (fixes second boot problem introduced by 0.7.0),
    • 0.7.4: Add rock(pro)64_erase_spi_flash.sh,
    • 0.7.4: Fix cursor on desktop for rockpro64,

    Bootet nur ohne PCIe NVMe Karte! Kamil will das Fixen.

  • 0.7.5 released

    • 0.7.5: Various stability fixes for kernel and u-boot,
    • 0.7.5: Added memtest to kernels and extlinux,
    • 0.7.5: Show early boot log when booting kernels,

    Kurzer Test - bootet jetzt mit PCIe NVMe Karte

  • 0.7.6 released

    • 0.7.6: Change OPP's for Rock64 and RockPro64: ayufan-rock64/linux-kernel@4.4.132-1059-rockchip-ayufan...ayufan-rock64:4.4.132-1062-rockchip-ayufan,

    Soll mehr Stabilität bringen, macht es leider nicht. So langsam wird es ärgerlich, das mit jedem Release nichts nach vorne geht.

  • 0.7.7 released

    • 0.7.7: Fix memory corruptions caused by Mali/Display subsystem (4.4),
    • 0.7.7: Enable SDR104 mode for SD cards (this requires u-boot upgrade if booting from SD),

    Langsam wird es, kann man jetzt zig mal booten ohne Probleme. Einziges Problem was ich im Moment nach einem kurzen Test festgestellt habe, das Erkennen der NVMe Karte ist Zufall und gelingt nur selten. Seltsamerweise im Mainline kein Problem!?

    Kamil hat das für 4.4 gefixt -> https://github.com/ayufan-rock64/linux-kernel/commit/bfb0d6c371d14b0d1fc60326b9bc84985a26f848?diff=unified

    Hier gibt es den Kernel zum herunter laden -> https://github.com/ayufan-rock64/linux-kernel/releases/tag/4.4.132-1070-rockchip-ayufan

  • 0.7.8 released ✌

    • 0.7.8: Improve eMMC compatibility on RockPro64,
    • 0.7.8: Disable sdio (no wifi/bt) to fix pcie/nvme support on 4.4 for RockPro64,
    • 0.7.8: Fix OMV builds (missing initrd.img),
    • 0.7.8: Make all packages virtual, conflicting and replacing making possible to do linux-rock64/rockpro64 to replace basesystem,

    (22:32:51) ayufan1: pushed one final release before vacations
    (22:32:59) ayufan1: consider this one to be the future release
    (22:33:15) ayufan1: if people do confirm that they do work on rockpro64 and rock64 I will make it finally stable 🙂
    (22:33:21) ayufan1: it is probably the best support so far
    (22:33:48) ayufan1: not yet the most performant (for rockpro64), I had to limit in bunch of places freqs, disable stuff, but it should be stable and support base ops
    (22:34:01) ayufan1: but, first stability, then we gonna add extra stuff

    Kamil macht Urlaub, da haben wir ja ein wenig Freizeit 🙂

  • 0.7.9 released

    • 0.7.9: Fix upgrade problem (u-boot-* packages),

    Kamil macht Urlaub 😂 Ok, es gab wohl ein paar kleinere Probleme die Platinen upzugraden. Kamil hat noch schnell die Probleme gefixt und das 0.7.9 Image veröffentlicht.

  • 0.7.10 released

    • 0.7.10: Rebased rockchip and mainline kernels,
    • 0.7.10: Support USB gadgets for rock/pro64,
    • 0.7.10: Disable TX checksumming for RockPro64,
    • 0.7.10: Improve FAN for RockPro64,
    • 0.7.10: Improve sdmmc0 stability for Rock64,
    • 0.7.10: Enable binfmt-misc,
    • 0.7.10: Improve stability of PCIE for RockPro64,
    • 0.7.10: Fix eMMC stability on RockPro64 mainline kernel,

    Die Container Images fehlen, Kamil hatte da ein Problem mit. Deshalb auch die Kennzeichnung als Pre-Release.

  • Die Container-Images für 0.7.10 sind jetzt auch fertig!

  • 0.7.11 released ✌

    • 0.7.11: Rebased mainline kernel,
    • 0.7.11: Run rockchip kernel at 250Hz to increase performance,
    • 0.7.11: Add support for usb gadgets for rockchip,
    • 0.7.11: Introduce change-default-kernel.sh script to easily switch between kernels,
  • 0.7.12 released

    • 0.7.12: Rebased mainline kernel,
    • 0.7.12: Rockchip kernel has patches for enabling sdio0 and pcie concurrently,
    • 0.7.12: A bunch of dependencies updates,
  • Ich kann 0.7.12 nicht empfehlen. Bei der Installation eines Systems mit PCIe NVMe SSD auf einem ROCKPro64 v2.0 4GB RAM habe ich folgenden Fehler gehabt.

    Danach war die Partition auf der SSD weg.

    Ich bin jetzt wieder auf 0.7.11 mit dem letzten Kernel der funktioniert 4.4.154-1128-rockchip-ayufan

    Dran denken, was ich mit diesem Zeichen ✌ kennzeichne, funktioniert.

  • 0.7.13 released

    • 0.7.13: Enable support for RockPro64 WiFi/BT module,
    • 0.7.13: Fix LXDE build: updated libdrm,

    PCIe NVMe SSD & WiFi = Crash!
    WiFi alleine habe ich bei einem kurzen Test auf stretch minimal zum Laufen bekommen - aber unstabil!

    Alles in allem, noch viel Arbeit. Nur für Leute mit Spaß am Testen geeignet 😉 Morgen, geht's weiter 🙂

  • 0.7.14 released

    • 0.7.14: Update rockchip kernel to 4.4.167,
    • 0.7.14: Update mainline kernel to 5.0,

    Aktuell noch ungetestet.

  • 0.7.14 ist nicht empfehlenswert. (Bei mir zumindestens!!)

    Was geht?

    • WiFi

    Was nicht geht?

    • WiFi & PCIe NVMe SSD zusammen - crash!
    • PCIe NVMe SSD alleine - crash!

    Für mich ist der angebliche Fix, der irgendwie was drumherum bastelt, damit PCIe & WiFi zusammen funktioniert keine Lösung. Das macht jetzt aktuell mehr Probleme als vorher.

    Meine Idee wäre, zwei dts Files, einmal ist WiFi an, für Leute die es brauchen und einmal ist es aus. Hoffentlich findet jemand eine vernünftige Lösung!!

    Ich bleibe weiterhin bei 0.7.11 - absolut stabil, PCIe stabil, SATA läuft (mit der richtigen Karte), auf WiFi kann ich persönlich verzichten!

    Ich möchte hier aber erwähnen, das es wohl Leute gibt, bei denen das funktioniert.

  • Mal ein 0.7.14 Image eingeschmissen und Mainline angetestet. Und jetzt wird es verwirrend..

    Was geht?

    • PCIe (auch mit gestecktem WiFi-Modul)

    Was nicht geht?

    • WiFi

  • Images 0.11.x

    Images
    5
    0 Stimmen
    5 Beiträge
    323 Aufrufe
    FrankMF

    0.11.2: gitlab-ci-linux-build-187 released

    0.11.2: Update OMV install (to also be able to run OMV6)
  • Buster & Disco Images in Arbeit

    Images
    3
    0 Stimmen
    3 Beiträge
    330 Aufrufe
    FrankMF

    Das passiert gerade, Images aber wohl noch nicht fertig.

    0.9.3: Update kernel branch to 4.4.185, 0.9.3: Disable OMV5 for now, 0.9.2: Fix linux-package release due to fpm being broken, 0.9.1: Build OMV5 (Debian/Buster), 0.9.1: Fix ppas for Debian/Buster and Ubuntu/Disco,
  • Desktop Image - endlich nutzbar??

    Images
    3
    0 Stimmen
    3 Beiträge
    318 Aufrufe
    FrankMF

    Bitte zum Thema WiFi & PCIe zusammen unbedingt hier mal reinschauen. Es gibt da interessante Neuigkeiten.

  • 0 Stimmen
    2 Beiträge
    487 Aufrufe
    FrankMF

    Das ist das Ergebnis des Stresstests 😞

    [ 2461.489468] ata2.00: exception Emask 0x10 SAct 0xffffffff SErr 0x400000 action 0x6 frozen [ 2461.490206] ata2.00: irq_stat 0x08000000, interface fatal error [ 2461.490732] ata2: SError: { Handshk } [ 2461.491062] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.491532] ata2.00: cmd 61/40:00:f8:a6:64/05:00:84:00:00/40 tag 0 ncq dma 688128 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.492993] ata2.00: status: { DRDY } [ 2461.493327] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.493796] ata2.00: cmd 61/c0:08:38:ac:64/03:00:84:00:00/40 tag 1 ncq dma 491520 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.495181] ata2.00: status: { DRDY } [ 2461.495507] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.496276] ata2.00: cmd 61/c0:10:a0:f5:64/02:00:84:00:00/40 tag 2 ncq dma 360448 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.497697] ata2.00: status: { DRDY } [ 2461.498029] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.498497] ata2.00: cmd 61/40:18:60:f8:64/05:00:84:00:00/40 tag 3 ncq dma 688128 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.499886] ata2.00: status: { DRDY } [ 2461.500213] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.500681] ata2.00: cmd 61/c0:20:a0:fd:64/02:00:84:00:00/40 tag 4 ncq dma 360448 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.502087] ata2.00: status: { DRDY } [ 2461.502416] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.502884] ata2.00: cmd 61/40:28:60:00:65/05:00:84:00:00/40 tag 5 ncq dma 688128 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.505026] ata2.00: status: { DRDY } [ 2461.505378] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.505852] ata2.00: cmd 61/c0:30:a0:05:65/02:00:84:00:00/40 tag 6 ncq dma 360448 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.507244] ata2.00: status: { DRDY } [ 2461.507572] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.508040] ata2.00: cmd 61/40:38:60:08:65/05:00:84:00:00/40 tag 7 ncq dma 688128 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.509472] ata2.00: status: { DRDY } [ 2461.509808] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.510277] ata2.00: cmd 61/a0:40:a0:0d:65/02:00:84:00:00/40 tag 8 ncq dma 344064 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.511667] ata2.00: status: { DRDY } [ 2461.511994] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.512461] ata2.00: cmd 61/40:48:00:20:66/05:00:84:00:00/40 tag 9 ncq dma 688128 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.514503] ata2.00: status: { DRDY } [ 2461.514850] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.515322] ata2.00: cmd 61/40:50:00:28:66/05:00:84:00:00/40 tag 10 ncq dma 688128 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.516721] ata2.00: status: { DRDY } [ 2461.517084] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.517562] ata2.00: cmd 61/c0:58:40:2d:66/02:00:84:00:00/40 tag 11 ncq dma 360448 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.519253] ata2.00: status: { DRDY } [ 2461.519595] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.520066] ata2.00: cmd 61/40:60:f8:9e:64/05:00:84:00:00/40 tag 12 ncq dma 688128 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.521504] ata2.00: status: { DRDY } [ 2461.521840] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.522309] ata2.00: cmd 61/c0:68:38:a4:64/02:00:84:00:00/40 tag 13 ncq dma 360448 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.523706] ata2.00: status: { DRDY } [ 2461.524033] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.524501] ata2.00: cmd 61/40:70:f8:af:64/05:00:84:00:00/40 tag 14 ncq dma 688128 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.525925] ata2.00: status: { DRDY } [ 2461.526256] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.526725] ata2.00: cmd 61/c0:78:38:b5:64/02:00:84:00:00/40 tag 15 ncq dma 360448 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.528122] ata2.00: status: { DRDY } [ 2461.528449] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.528939] ata2.00: cmd 61/40:80:f8:b7:64/05:00:84:00:00/40 tag 16 ncq dma 688128 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.530339] ata2.00: status: { DRDY } [ 2461.530667] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.531136] ata2.00: cmd 61/c0:88:38:bd:64/02:00:84:00:00/40 tag 17 ncq dma 360448 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.532532] ata2.00: status: { DRDY } [ 2461.532880] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.533357] ata2.00: cmd 61/40:90:f8:bf:64/05:00:84:00:00/40 tag 18 ncq dma 688128 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.534754] ata2.00: status: { DRDY } [ 2461.535081] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.535549] ata2.00: cmd 61/c0:98:38:c5:64/02:00:84:00:00/40 tag 19 ncq dma 360448 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.536970] ata2.00: status: { DRDY } [ 2461.537301] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.537769] ata2.00: cmd 61/40:a0:f8:c7:64/05:00:84:00:00/40 tag 20 ncq dma 688128 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.539165] ata2.00: status: { DRDY } [ 2461.539491] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.539960] ata2.00: cmd 61/c0:a8:38:cd:64/02:00:84:00:00/40 tag 21 ncq dma 360448 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.541381] ata2.00: status: { DRDY } [ 2461.541713] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.542182] ata2.00: cmd 61/40:b0:f8:cf:64/05:00:84:00:00/40 tag 22 ncq dma 688128 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.543577] ata2.00: status: { DRDY } [ 2461.543905] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.544374] ata2.00: cmd 61/40:b8:f8:d7:64/05:00:84:00:00/40 tag 23 ncq dma 688128 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.545790] ata2.00: status: { DRDY } [ 2461.546120] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.546589] ata2.00: cmd 61/c0:c0:38:dd:64/02:00:84:00:00/40 tag 24 ncq dma 360448 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.547987] ata2.00: status: { DRDY } [ 2461.548314] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.548782] ata2.00: cmd 61/a8:c8:f8:df:64/05:00:84:00:00/40 tag 25 ncq dma 741376 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.550198] ata2.00: status: { DRDY } [ 2461.550530] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.550999] ata2.00: cmd 61/c0:d0:a0:e5:64/02:00:84:00:00/40 tag 26 ncq dma 360448 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.552396] ata2.00: status: { DRDY } [ 2461.552723] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.553208] ata2.00: cmd 61/40:d8:60:e8:64/05:00:84:00:00/40 tag 27 ncq dma 688128 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.554607] ata2.00: status: { DRDY } [ 2461.554935] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.555404] ata2.00: cmd 61/c0:e0:a0:ed:64/02:00:84:00:00/40 tag 28 ncq dma 360448 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.556800] ata2.00: status: { DRDY } [ 2461.557145] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.557617] ata2.00: cmd 61/40:e8:60:f0:64/05:00:84:00:00/40 tag 29 ncq dma 688128 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.559012] ata2.00: status: { DRDY } [ 2461.559340] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.559807] ata2.00: cmd 61/c0:f0:38:d5:64/02:00:84:00:00/40 tag 30 ncq dma 360448 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.561221] ata2.00: status: { DRDY } [ 2461.561552] ata2.00: failed command: WRITE FPDMA QUEUED [ 2461.562021] ata2.00: cmd 61/c0:f8:40:25:66/02:00:84:00:00/40 tag 31 ncq dma 360448 out res 40/00:50:00:28:66/00:00:84:00:00/40 Emask 0x10 (ATA bus error) [ 2461.563416] ata2.00: status: { DRDY } [ 2461.563752] ata2: hard resetting link [ 2471.561504] ata2: softreset failed (1st FIS failed) [ 2471.561959] ata2: hard resetting link [ 2481.560785] ata2: softreset failed (1st FIS failed) [ 2481.561238] ata2: hard resetting link [ 2516.561654] ata2: softreset failed (1st FIS failed) [ 2516.562109] ata2: limiting SATA link speed to 3.0 Gbps [ 2516.562113] ata2: hard resetting link [ 2521.561261] ata2: softreset failed (1st FIS failed) [ 2521.561715] ata2: reset failed, giving up [ 2521.562074] ata2.00: disabled [ 2521.562575] ata2: EH complete [ 2521.562677] sd 1:0:0:0: [sdb] tag#18 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2521.562686] sd 1:0:0:0: [sdb] tag#18 CDB: opcode=0x35 35 00 00 00 00 00 00 00 00 00 [ 2521.562701] print_req_error: I/O error, dev sdb, sector 0 [ 2521.562791] sd 1:0:0:0: [sdb] tag#20 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2521.562802] sd 1:0:0:0: [sdb] tag#19 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2521.562873] sd 1:0:0:0: [sdb] tag#19 CDB: opcode=0x2a 2a 00 84 66 25 40 00 02 c0 00 [ 2521.562888] print_req_error: I/O error, dev sdb, sector 2221286720 [ 2521.563107] sd 1:0:0:0: [sdb] tag#21 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2521.563126] sd 1:0:0:0: [sdb] tag#21 CDB: opcode=0x2a 2a 00 84 64 f0 60 00 05 40 00 [ 2521.563138] print_req_error: I/O error, dev sdb, sector 2221207648 [ 2521.563422] sd 1:0:0:0: [sdb] tag#22 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2521.563772] sd 1:0:0:0: [sdb] tag#20 CDB: opcode=0x2a 2a 00 84 64 d5 38 00 02 c0 00 [ 2521.564326] print_req_error: I/O error, dev sdb, sector 2221200696 [ 2521.564337] sd 1:0:0:0: [sdb] tag#22 CDB: opcode=0x2a 2a 00 84 64 ed a0 00 02 c0 00 [ 2521.564896] print_req_error: I/O error, dev sdb, sector 2221206944 [ 2521.565088] sd 1:0:0:0: [sdb] tag#26 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2521.565469] sd 1:0:0:0: [sdb] tag#26 CDB: opcode=0x2a 2a 00 84 64 dd 38 00 02 c0 00 [ 2521.565483] print_req_error: I/O error, dev sdb, sector 2221202744 [ 2521.565610] sd 1:0:0:0: [sdb] tag#23 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2521.566056] sd 1:0:0:0: [sdb] tag#23 CDB: opcode=0x2a 2a 00 84 64 e8 60 00 05 40 00 [ 2521.566069] print_req_error: I/O error, dev sdb, sector 2221205600 [ 2521.566183] sd 1:0:0:0: [sdb] tag#29 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2521.566641] sd 1:0:0:0: [sdb] tag#29 CDB: opcode=0x2a 2a 00 84 64 cd 38 00 02 c0 00 [ 2521.566654] print_req_error: I/O error, dev sdb, sector 2221198648 [ 2521.566954] sd 1:0:0:0: [sdb] tag#24 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2521.567224] sd 1:0:0:0: [sdb] tag#24 CDB: opcode=0x2a 2a 00 84 64 e5 a0 00 02 c0 00 [ 2521.567237] print_req_error: I/O error, dev sdb, sector 2221204896 [ 2521.567459] sd 1:0:0:0: [sdb] tag#30 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2521.567809] sd 1:0:0:0: [sdb] tag#30 CDB: opcode=0x2a 2a 00 84 64 c7 f8 00 05 40 00 [ 2521.567821] print_req_error: I/O error, dev sdb, sector 2221197304 [ 2521.584903] md: super_written gets error=10 [ 2521.585306] md/raid1:md0: Disk failure on dm-1, disabling device. md/raid1:md0: Operation continuing on 1 devices. [ 2526.581450] scsi_io_completion_action: 41779 callbacks suppressed [ 2526.581461] sd 1:0:0:0: [sdb] tag#27 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2526.581467] sd 1:0:0:0: [sdb] tag#27 CDB: opcode=0x2a 2a 00 84 66 7d 50 00 00 01 00 [ 2526.581469] print_req_error: 41780 callbacks suppressed [ 2526.581471] print_req_error: I/O error, dev sdb, sector 2221309264 [ 2526.582524] sd 1:0:0:0: [sdb] tag#28 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2526.582530] sd 1:0:0:0: [sdb] tag#28 CDB: opcode=0x2a 2a 00 84 66 7d 51 00 00 01 00 [ 2526.582534] print_req_error: I/O error, dev sdb, sector 2221309265 [ 2526.583366] sd 1:0:0:0: [sdb] tag#29 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2526.583370] sd 1:0:0:0: [sdb] tag#29 CDB: opcode=0x2a 2a 00 84 66 7d 52 00 00 01 00 [ 2526.583373] print_req_error: I/O error, dev sdb, sector 2221309266 [ 2526.584113] sd 1:0:0:0: [sdb] tag#30 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2526.584117] sd 1:0:0:0: [sdb] tag#30 CDB: opcode=0x2a 2a 00 84 66 7d 53 00 00 01 00 [ 2526.584119] print_req_error: I/O error, dev sdb, sector 2221309267 [ 2526.584960] sd 1:0:0:0: [sdb] tag#0 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2526.584968] sd 1:0:0:0: [sdb] tag#0 CDB: opcode=0x2a 2a 00 84 66 7d 54 00 00 01 00 [ 2526.584971] print_req_error: I/O error, dev sdb, sector 2221309268 [ 2526.585765] sd 1:0:0:0: [sdb] tag#1 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2526.585769] sd 1:0:0:0: [sdb] tag#1 CDB: opcode=0x2a 2a 00 84 66 7d 55 00 00 01 00 [ 2526.585772] print_req_error: I/O error, dev sdb, sector 2221309269 [ 2526.586461] sd 1:0:0:0: [sdb] tag#2 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2526.586465] sd 1:0:0:0: [sdb] tag#2 CDB: opcode=0x2a 2a 00 84 66 7d 56 00 00 01 00 [ 2526.586468] print_req_error: I/O error, dev sdb, sector 2221309270 [ 2526.587144] sd 1:0:0:0: [sdb] tag#3 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2526.587148] sd 1:0:0:0: [sdb] tag#3 CDB: opcode=0x2a 2a 00 84 66 7d 57 00 00 01 00 [ 2526.587150] print_req_error: I/O error, dev sdb, sector 2221309271 [ 2526.587734] sd 1:0:0:0: [sdb] tag#4 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2526.587737] sd 1:0:0:0: [sdb] tag#4 CDB: opcode=0x2a 2a 00 84 66 7d 58 00 00 01 00 [ 2526.587739] print_req_error: I/O error, dev sdb, sector 2221309272 [ 2526.588320] sd 1:0:0:0: [sdb] tag#5 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00 [ 2526.588323] sd 1:0:0:0: [sdb] tag#5 CDB: opcode=0x2a 2a 00 84 66 7d 59 00 00 01 00 [ 2526.588325] print_req_error: I/O error, dev sdb, sector 2221309273

    Und, wieder eine Platte im Raid1 verloren

    rock64@rockpro64v_2_1:~$ cat /proc/mdstat Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10] md0 : active raid1 dm-1[2](F) dm-0[1] 1953379392 blocks super 1.2 [2/1] [_U] bitmap: 5/15 pages [20KB], 65536KB chunk unused devices: <none>

    Leider habe ich nicht die Kenntnisse um zu erkennen, woran das liegt.

  • Batocera Image released

    Images
    1
    0 Stimmen
    1 Beiträge
    268 Aufrufe
    Niemand hat geantwortet
  • ROCKPro64 - Wo stehen wir?

    ROCKPro64
    1
    0 Stimmen
    1 Beiträge
    581 Aufrufe
    Niemand hat geantwortet
  • stretch-minimal-rockpro64

    Verschoben Linux
    3
    0 Stimmen
    3 Beiträge
    983 Aufrufe
    FrankMF

    Mal ein Test was der Speicher so kann.

    rock64@rockpro64:~/tinymembench$ ./tinymembench tinymembench v0.4.9 (simple benchmark for memory throughput and latency) ========================================================================== == Memory bandwidth tests == == == == Note 1: 1MB = 1000000 bytes == == Note 2: Results for 'copy' tests show how many bytes can be == == copied per second (adding together read and writen == == bytes would have provided twice higher numbers) == == Note 3: 2-pass copy means that we are using a small temporary buffer == == to first fetch data into it, and only then write it to the == == destination (source -> L1 cache, L1 cache -> destination) == == Note 4: If sample standard deviation exceeds 0.1%, it is shown in == == brackets == ========================================================================== C copy backwards : 2812.7 MB/s C copy backwards (32 byte blocks) : 2811.9 MB/s C copy backwards (64 byte blocks) : 2632.8 MB/s C copy : 2667.2 MB/s C copy prefetched (32 bytes step) : 2633.5 MB/s C copy prefetched (64 bytes step) : 2640.8 MB/s C 2-pass copy : 2509.8 MB/s C 2-pass copy prefetched (32 bytes step) : 2431.6 MB/s C 2-pass copy prefetched (64 bytes step) : 2424.1 MB/s C fill : 4887.7 MB/s (0.5%) C fill (shuffle within 16 byte blocks) : 4883.0 MB/s C fill (shuffle within 32 byte blocks) : 4889.3 MB/s C fill (shuffle within 64 byte blocks) : 4889.2 MB/s --- standard memcpy : 2807.3 MB/s standard memset : 4890.4 MB/s (0.3%) --- NEON LDP/STP copy : 2803.7 MB/s NEON LDP/STP copy pldl2strm (32 bytes step) : 2802.1 MB/s NEON LDP/STP copy pldl2strm (64 bytes step) : 2800.7 MB/s NEON LDP/STP copy pldl1keep (32 bytes step) : 2745.5 MB/s NEON LDP/STP copy pldl1keep (64 bytes step) : 2745.8 MB/s NEON LD1/ST1 copy : 2801.9 MB/s NEON STP fill : 4888.9 MB/s (0.3%) NEON STNP fill : 4850.1 MB/s ARM LDP/STP copy : 2803.8 MB/s ARM STP fill : 4893.0 MB/s (0.5%) ARM STNP fill : 4851.7 MB/s ========================================================================== == Framebuffer read tests. == == == == Many ARM devices use a part of the system memory as the framebuffer, == == typically mapped as uncached but with write-combining enabled. == == Writes to such framebuffers are quite fast, but reads are much == == slower and very sensitive to the alignment and the selection of == == CPU instructions which are used for accessing memory. == == == == Many x86 systems allocate the framebuffer in the GPU memory, == == accessible for the CPU via a relatively slow PCI-E bus. Moreover, == == PCI-E is asymmetric and handles reads a lot worse than writes. == == == == If uncached framebuffer reads are reasonably fast (at least 100 MB/s == == or preferably >300 MB/s), then using the shadow framebuffer layer == == is not necessary in Xorg DDX drivers, resulting in a nice overall == == performance improvement. For example, the xf86-video-fbturbo DDX == == uses this trick. == ========================================================================== NEON LDP/STP copy (from framebuffer) : 602.5 MB/s NEON LDP/STP 2-pass copy (from framebuffer) : 551.6 MB/s NEON LD1/ST1 copy (from framebuffer) : 667.1 MB/s NEON LD1/ST1 2-pass copy (from framebuffer) : 605.6 MB/s ARM LDP/STP copy (from framebuffer) : 445.3 MB/s ARM LDP/STP 2-pass copy (from framebuffer) : 428.8 MB/s ========================================================================== == Memory latency test == == == == Average time is measured for random memory accesses in the buffers == == of different sizes. The larger is the buffer, the more significant == == are relative contributions of TLB, L1/L2 cache misses and SDRAM == == accesses. For extremely large buffer sizes we are expecting to see == == page table walk with several requests to SDRAM for almost every == == memory access (though 64MiB is not nearly large enough to experience == == this effect to its fullest). == == == == Note 1: All the numbers are representing extra time, which needs to == == be added to L1 cache latency. The cycle timings for L1 cache == == latency can be usually found in the processor documentation. == == Note 2: Dual random read means that we are simultaneously performing == == two independent memory accesses at a time. In the case if == == the memory subsystem can't handle multiple outstanding == == requests, dual random read has the same timings as two == == single reads performed one after another. == ========================================================================== block size : single random read / dual random read 1024 : 0.0 ns / 0.0 ns 2048 : 0.0 ns / 0.0 ns 4096 : 0.0 ns / 0.0 ns 8192 : 0.0 ns / 0.0 ns 16384 : 0.0 ns / 0.0 ns 32768 : 0.0 ns / 0.0 ns 65536 : 4.5 ns / 7.2 ns 131072 : 6.8 ns / 9.7 ns 262144 : 9.8 ns / 12.8 ns 524288 : 11.4 ns / 14.7 ns 1048576 : 16.0 ns / 22.6 ns 2097152 : 114.0 ns / 175.3 ns 4194304 : 161.7 ns / 219.9 ns 8388608 : 190.7 ns / 241.5 ns 16777216 : 205.3 ns / 250.5 ns 33554432 : 212.9 ns / 255.5 ns 67108864 : 222.3 ns / 271.1 ns
  • 0 Stimmen
    5 Beiträge
    1k Aufrufe
    FrankMF

    Neue Version

    Stock for install from SDcard to eMMC [microSD Boot] [20180921]

    Link Preview Image ROCKPro64 Software Releases

    favicon

    PINE64 (wiki.pine64.org)