Store 2.0: Unterschied zwischen den Versionen

Aus Store2 Wiki
Zur Navigation springen Zur Suche springen
Zeile 699: Zeile 699:


Versuch von --update=revert-reshape
Versuch von --update=revert-reshape
--assemble --update=revert-reshape" can be used to undo a reshape
--assemble --update=revert-reshape" can be used to undo a reshape
   that has just been started but isn't really wanted.  This is very
   that has just been started but isn't really wanted.  This is very
   new and while it passes basic tests it cannot be guaranteed.
   new and while it passes basic tests it cannot be guaranteed.

Version vom 13. September 2015, 07:49 Uhr

STABILITÄT Store 2.0

Vermutung: Die drei SATA-Erweiterungskarten crashen gelegentlich das komplette System.

Karten:

http://www.sybausa.com/productInfo.php?iid=537 

Syba SY-PEX40008 4-port SATA II PCI-e Software RAID Controller Card--Bundle with Low Profile Bracket, SIL3124 Chipset Sind die identischen Karten, die immer noch von Backblaze verbaut werden (Pod 2.0 UND Pod 3.0!) Hängen an drei PCI-E 1x (kleiner Port)

RAID bliebt heile (zum Glück!), da die Karten dann komplett die Zugriff sperren.

Arch-Log files haben gar keine Einträge zu den Crashs!!!

Remote-Log über Syslogd (über Myth) zeigt als letzten Eintrag:

 mdadm: sending ioctl 1261 to a partition (buggy Eintrag, aber unkritisch)
 sata_sil24: IRQ status == 0xffffffff, PCI fault or device removal

sata-sil24:

https://ata.wiki.kernel.org/index.php/Sata_sil24 

Spurious interrupts are expected on SiI3124 suffering from IRQ loss erratum on PCI-X


PATCH?

http://old.nabble.com/-PATCH-06-13--sata_sil24%3A-implement-loss-of-completion-interrupt-on-PCI-X-errta-fix-p3799674.html

Thread über Zugang mit SIL3124 Chip

http://www.linuxquestions.org/questions/linux-kernel-70/how-to-access-sata-drives-attached-to-sii3124-719408/

Test?

http://marc.info/?l=linux-ide&m=127228317404771&w=2

Raid nach Booten öffnen und mount

Um Auto-Assembly beim Booten zu verhindern muss die Config-Datei /etc/mdadm.conf leer (oder zumindest komplett auskommentiert sein) und "MDADM_SCAN=no" in /etc/sysconfig/mdadm

1.) Checken ob alle Platten da sind:

/root/bin/diskserial_sort2.sh 

Müssen im Moment 17 Platten sein. Basis ist die Datei disknum.txt unter /root/bin

2.) Raids suchen und assemblen (kein Autostart):

mdadm --assemble --scan 

3.) Cryptsetup:

cryptsetup luksOpen /dev/md125 cr_md125 

4.) Mounten:

mount /dev/mapper/cr_md125 /data

Schliessen wäre:

cryptsetup luksClose cr_md125

JD2

java -Xmx512m -jar /home/gagi/jd2/JDownloader.jar

bzw in /home/gagi/jd2/

./JDownloader2

VNC

dergagi.selfhost.bz:5901

Festplatten-Layout

3000GB Hitachi Deskstar 5K3000 HDS5C3030ALA630 CoolSpin 32MB 3.5" (8.9cm) SATA 6Gb/s

3000GB Western Digital WD30EZRX 3TB interne Festplatte (8,9 cm (3,5 Zoll), 5400 rpm, 2ms, 64MB Cache, SATA III

Problem mit WD-Platten und LCC

http://idle3-tools.sourceforge.net/ 
http://koitsu.wordpress.com/2012/05/30/wd30ezrx-and-aggressive-head-parking/ 

Get idle3 timer raw value

idle3ctl -g /dev/sdh 

Disable idle3 timer:

idle3ctl -d /dev/sdh 

Serial auslesen mit:

udevadm info --query=all --name=/dev/sdi | grep ID_SERIAL_SHORT 

Serial Systemplatte 160GB:

JC0150HT0J7TPC

Serials der Datenplatten

00  : 00000000000000 (1TB System, Samsung)

geht

01  : 
02  : 
03  : 
04  :
05  : 

geht jetzt auch, Molex-Kontakt Problem behoben

06  : 
07  : MJ1311YNG4J48A (3TB)
08  : WD-WCC070299387 (3TB WD)
09  : MJ1311YNG3UUPA (3TB)
10  : 

geht

11  : MJ1311YNG3SAMA (3TB)
12  : 13V9WK9AS NEU Hot Spare data
13  : MJ1311YNG09EDA (3TB) Garantie-Austausch, HOT SPARE data2
14  : 
15  : MCE9215Q0AUYTW (3TB Toshiba neu)

geht

16  : MJ0351YNGA02YA (3TB)
17  : 
18  : 
19  : 
20  : WD-WCAWZ1881335 (3TB WD)

geht jetzt auch, Molex-Kontakt Problem behoben

21  : 
22  : WD-WCAWZ2279670 (3TB WD)
23  : MJ1311YNG3SSLA (3TB)
24  : MJ1311YNG25Z6A (3TB)
25  : 

geht

26  : MJ1311YNG3RM5A (3TB)
27  : MJ1311YNG3NZ3A (3TB)
28  : MJ1311YNG3NT5A (3TB)
29  : 
30  : MCM9215Q0B9LSY (3TB Toshiba neu)

geht

31  : 234BGY0GS (3TB Toshiba neu)
32  : 
33  : MJ1311YNG3WZVA (3TB)
34  : MJ1311YNG3Y4SA (3TB)
35  : MJ1311YNG3SYKA (3TB)

geht

36  : 
37  : WD-WCC070198169 (3TB WD)
38  : 
39  : MJ1311YNG3RZTA (3TB)
40  : 

geht

41  : MJ1311YNG3LTRA (3TB)
42  : 
43  : MJ1311YNG38VGA (3TB)
44  :
45  :

TOTAL: 25 (18 + 4 + 2 Hotspare + eine Systemplatte) von 46 möglichen

Raid Baubefehl

im Screen mdadm

mdadm --create /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=15 /dev/sdb1 /dev/sdc1 /dev/sdd1 /dev/sde1 /dev/sdf1 /dev/sdg1 /dev/sdh1 /dev/sdi1 /dev/sdj1 /dev/sdk1 /dev/sdl1 /dev/sdm1 /dev/sdn1 /dev/sdo1 /dev/sdp1


Re-Create 2014-01-31: NEUER RICHTIGER RE-CREATE BEFEHL von mdadm-git:

./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdb1:1024 /dev/sdd1:1024 /dev/sdf1:1024 /dev/sdg1:1024 /dev/sdh1:1024 /dev/sdc1:1024 /dev/sdt1:1024 /dev/sdn1:1024 /dev/sdo1:1024 /dev/sdq1:1024 /dev/sdm1:1024 /dev/sdp1:1024 /dev/sdu1:1024 /dev/sdv1:1024 /dev/sda1:1024 /dev/sds1:1024 /dev/sdl1:1024 /dev/sdw1:1024

Re-Create 2014-05-23: NEUER RICHTIGER RE-CREATE BEFEHL von mdadm-git aus /builds/mdadm/:

./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdb1:1024 /dev/sdd1:1024 /dev/sdf1:1024 /dev/sdg1:1024 /dev/sdh1:1024 /dev/sdc1:1024 /dev/sdt1:1024 /dev/sdm1:1024 /dev/sdn1:1024 /dev/sdp1:1024 /dev/sdl1:1024 /dev/sdo1:1024 /dev/sdu1:1024 /dev/sdv1:1024 /dev/sda1:1024 /dev/sds1:1024 /dev/sdk1:1024 /dev/sdw1:1024

Re-Create 2014-07-10: NEUER RICHTIGER RE-CREATE BEFEHL von mdadm-git aus /builds/mdadm/:

./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdb1:1024 /dev/sdd1:1024 /dev/sdf1:1024 /dev/sdg1:1024 /dev/sdh1:1024 /dev/sdc1:1024 /dev/sdt1:1024 /dev/sdn1:1024 /dev/sdo1:1024 /dev/sdq1:1024 /dev/sdm1:1024 /dev/sdp1:1024 /dev/sdu1:1024 /dev/sdv1:1024 /dev/sda1:1024 /dev/sds1:1024 /dev/sdl1:1024 /dev/sdw1:1024

Zweites Raid Baubefehl DATA2

im Screen mdadm

mdadm --create /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=4 /dev/sdv1 /dev/sdw1 /dev/sdj1 /dev/sdh1

Verschlüsseln mit speziellen Paramtern für Hardware-Verschlüsselung:

cryptsetup -v luksFormat --cipher aes-cbc-essiv:sha256 --key-size 256 /dev/md126

Öffnen:

cryptsetup luksOpen /dev/md126 cr_md126 

XFS Filesystem drauf:

mkfs.xfs /dev/mapper/cr_md126

Versuch Recreate 2014-10-12 nach missglücktem Grow von 4 auf 5 RAID-Devices, scheiße:

./mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=4 /dev/sdy1:1024 /dev/sdq1:1024 /dev/sdi1:1024 /dev/sdj1:1024

Echtes Recreate 2014-10-12:

mdadm --create /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdq1 /dev/sdi1 /dev/sdj1 /dev/sdx1

Drittes Raid Recreate DATA und DATA2 2015-01-20

./mdadm --create /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1:1024 /dev/sdr1:1024 /dev/sdj1:1024 /dev/sdk1:1024 /dev/sdx1:1024
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdm1:1024 /dev/sdo1:1024 /dev/sdi1:1024 /dev/sdb1:1024 /dev/sdp1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdf1:1024 /dev/sdd1:1024 /dev/sdh1:1024 /dev/sde1:1024 /dev/sdg1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdk1:1024 /dev/sdq1:1024 /dev/sdc1:1024 /dev/sdv1:1024

Builden OHNE Spare, also 18 Devices, dann Spare adden

Viertes Raid Recreate DATA 2015-02-01

in /root/bin

./diskserial2.sh 

um die Festplattenreihenfolge im Raid zu sehen. Ohne hddtemp etc. wohl auch stabiler

mdadm aus /builds/mdadm

./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdo1:1024 /dev/sdd1:1024 /dev/sdk1:1024 /dev/sdg1:1024 /dev/sdp1:1024 /dev/sdb1:1024 /dev/sdr1:1024 /dev/sdf1:1024 /dev/sdh1:1024 /dev/sdj1:1024 /dev/sde1:1024 /dev/sdg1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdm1:1024 /dev/sdq1:1024 /dev/sdc1:1024 /dev/sdv1:1024 


Mit normalem mdadm gehts wohl auch

mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdo1 /dev/sdd1 /dev/sdk1 /dev/sdg1 /dev/sdp1 /dev/sdb1 /dev/sdr1 /dev/sdf1 /dev/sdh1 /dev/sdj1 /dev/sde1 /dev/sdg1 /dev/sds1 /dev/sdu1 /dev/sdm1 /dev/sdq1 /dev/sdc1 /dev/sdv1 

Builden OHNE Spare, also 18 Devices, dann Spare adden

PROBLEM:

mdadm: failed to open /dev/sdg1 after earlier success - aborting

Fünftes Raid Recreate DATA 2015-02-28

in /root/bin

./diskserial2.sh 

um die Festplattenreihenfolge im Raid zu sehen.

mdadm aus /builds/mdadm

dqelcpfhmakgsunrjv 
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdd1:1024 /dev/sdq1:1024 /dev/sde1:1024 /dev/sdl1:1024 /dev/sdc1:1024 /dev/sdp1:1024 /dev/sdf1:1024 /dev/sdh1:1024 /dev/sdm1:1024 /dev/sda1:1024 /dev/sdk1:1024 /dev/sdg1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdj1:1024 /dev/sdv1:1024 

Builden OHNE Spare, also 18 Devices, dann Spare adden

Sechstes Raid Recreate DATA und DATA2 2015-03-11

3 neue aber identische Sil-Sata Raid PCI1x Karten mit Firmware 2.6.18

in /root/bin

./diskserial2.sh 

um die Festplattenreihenfolge im Raid zu sehen.

DATA2 mit normalem mdadm

ybtxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdb1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

dqemcpfhialgsunrkv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdd1:1024 /dev/sdq1:1024 /dev/sde1:1024 /dev/sdm1:1024 /dev/sdc1:1024 /dev/sdp1:1024 /dev/sdf1:1024 /dev/sdh1:1024 /dev/sdi1:1024 /dev/sda1:1024 /dev/sdl1:1024 /dev/sdg1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdk1:1024 /dev/sdv1:1024 

Builden OHNE Spare, also 18 Devices, dann Spare adden

Siebtes Raid Recreate DATA und DATA2 2015-03-15

3 neue aber identische Sil-Sata Raid PCI1x Karten mit Firmware 2.6.18

in /root/bin

./diskserial2.sh 

um die Festplattenreihenfolge im Raid zu sehen.

DATA2 mit normalem mdadm

ybixw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdb1 /dev/sdi1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

eqjdcprtgalshunfkv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sde1:1024 /dev/sdq1:1024 /dev/sdj1:1024 /dev/sdd1:1024 /dev/sdc1:1024 /dev/sdp1:1024 /dev/sdr1:1024 /dev/sdt1:1024 /dev/sdg1:1024 /dev/sda1:1024 /dev/sdl1:1024 /dev/sds1:1024 /dev/sdh1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdf1:1024 /dev/sdk1:1024 /dev/sdv1:1024 

Builden OHNE Spare, also 18 Devices, dann Spare adden

Achtes Raid Recreate DATA und DATA2 2015-03-21

DATA2 mit normalem mdadm

ybtxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdb1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

eqfldpgicakhsunrjv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sde1:1024 /dev/sdq1:1024 /dev/sdf1:1024 /dev/sdl1:1024 /dev/sdd1:1024 /dev/sdp1:1024 /dev/sdg1:1024 /dev/sdi1:1024 /dev/sdc1:1024 /dev/sda1:1024 /dev/sdk1:1024 /dev/sdh1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdj1:1024 /dev/sdv1:1024

Neuntes Raid Recreate DATA und DATA2 2015-04-22

DATA2 mit normalem mdadm

yctxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdc1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

eqfldpgimbkhsunrjv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sde1:1024 /dev/sdq1:1024 /dev/sdf1:1024 /dev/sdl1:1024 /dev/sdd1:1024 /dev/sdp1:1024 /dev/sdg1:1024 /dev/sdi1:1024 /dev/sdm1:1024 /dev/sdb1:1024 /dev/sdk1:1024 /dev/sdh1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdj1:1024 /dev/sdv1:1024

Zehntes Raid Recreate DATA und DATA2 2015-04-24

DATA2 mit normalem mdadm

yctxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdc1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

eqfldpgimbkhsunrjv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sde1:1024 /dev/sdq1:1024 /dev/sdf1:1024 /dev/sdl1:1024 /dev/sdd1:1024 /dev/sdp1:1024 /dev/sdg1:1024 /dev/sdi1:1024 /dev/sdm1:1024 /dev/sdb1:1024 /dev/sdk1:1024 /dev/sdh1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdj1:1024 /dev/sdv1:1024

Elftes Raid Recreate DATA und DATA2 2015-04-26

DATA2 mit normalem mdadm

ybtxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdb1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

eqfmcpgidalhsunrkv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sde1:1024 /dev/sdq1:1024 /dev/sdf1:1024 /dev/sdm1:1024 /dev/sdc1:1024 /dev/sdp1:1024 /dev/sdg1:1024 /dev/sdi1:1024 /dev/sdd1:1024 /dev/sda1:1024 /dev/sdl1:1024 /dev/sdh1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdk1:1024 /dev/sdv1:1024

12. Raid Recreate DATA und DATA2 2015-05-06

DATA2 mit normalem mdadm

ybtxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdb1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

dqel cpfh makg sunrjv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdd1:1024 /dev/sdq1:1024 /dev/sde1:1024 /dev/sdl1:1024 /dev/sdc1:1024 /dev/sdp1:1024 /dev/sdf1:1024 /dev/sdh1:1024 /dev/sdm1:1024 /dev/sda1:1024 /dev/sdk1:1024 /dev/sdg1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdj1:1024 /dev/sdv1:1024


smartctl-timeouts ergänzt 2015-05-07

/etc/udev/rules.d/ 
http://article.gmane.org/gmane.linux.raid/48238/match=smartctl+timeouts

The smartctl-timeouts scripts fix commonly mismatching defaults with drives that have no error recovery timeout configured, which has often lead to data loss.

To test, extract the files to /etc/udev/rules.d/ and

  • reboot*. For me, the rules somehow had no effect without rebooting.

13. Raid Recreate DATA und DATA2 2015-05-11

DATA2 mit normalem mdadm

xcswv
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdx1 /dev/sdc1 /dev/sds1 /dev/sdw1 /dev/sdv1 

DATA mdadm aus /builds/mdadm

epfl dogi bakh rtmqju
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sde1:1024 /dev/sdp1:1024 /dev/sdf1:1024 /dev/sdl1:1024 /dev/sdd1:1024 /dev/sdo1:1024 /dev/sdg1:1024 /dev/sdi1:1024 /dev/sdb1:1024 /dev/sda1:1024 /dev/sdk1:1024 /dev/sdh1:1024 /dev/sdr1:1024 /dev/sdt1:1024 /dev/sdm1:1024 /dev/sdq1:1024 /dev/sdj1:1024 /dev/sdu1:1024

14. Raid Recreate DATA und DATA2 2015-05-14

DATA2 mit normalem mdadm

xbswv
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdx1 /dev/sdb1 /dev/sds1 /dev/sdw1 /dev/sdv1 

DATA mdadm aus /builds/mdadm

dpek cofh lajg rtmqiu
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdd1:1024 /dev/sdp1:1024 /dev/sde1:1024 /dev/sdk1:1024 /dev/sdc1:1024 /dev/sdo1:1024 /dev/sdf1:1024 /dev/sdh1:1024 /dev/sdl1:1024 /dev/sda1:1024 /dev/sdj1:1024 /dev/sdg1:1024 /dev/sdr1:1024 /dev/sdt1:1024 /dev/sdm1:1024 /dev/sdq1:1024 /dev/sdi1:1024 /dev/sdu1:1024

15. Raid Recreate DATA und DATA2 2015-05-20

DATA2 mit normalem mdadm

yctxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdc1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

eqfl dpgi mbkh sunrjv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sde1:1024 /dev/sdq1:1024 /dev/sdf1:1024 /dev/sdl1:1024 /dev/sdd1:1024 /dev/sdp1:1024 /dev/sdg1:1024 /dev/sdi1:1024 /dev/sdm1:1024 /dev/sdb1:1024 /dev/sdk1:1024 /dev/sdh1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdj1:1024 /dev/sdv1:1024

16. Raid Recreate DATA und DATA2 2015-05-21

DATA2 mit normalem mdadm

yctxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdc1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

eqfb 4 dpgi 8 jalh 12 sunrkv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sde1:1024 /dev/sdq1:1024 /dev/sdf1:1024 /dev/sdb1:1024 /dev/sdd1:1024 /dev/sdp1:1024 /dev/sdg1:1024 /dev/sdi1:1024 /dev/sdj1:1024 /dev/sda1:1024 /dev/sdl1:1024 /dev/sdh1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdk1:1024 /dev/sdv1:1024

17. Raid Recreate DATA und DATA2 2015-05-22

DATA2 mit normalem mdadm

ydtxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdd1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

cqfl 4 epgi 8 mbkh 12 sunrjv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdc1:1024 /dev/sdq1:1024 /dev/sdf1:1024 /dev/sdl1:1024 /dev/sde1:1024 /dev/sdp1:1024 /dev/sdg1:1024 /dev/sdi1:1024 /dev/sdm1:1024 /dev/sdb1:1024 /dev/sdk1:1024 /dev/sdh1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdj1:1024 /dev/sdv1:1024


18. Raid Recreate DATA und DATA2 2015-05-23

DATA2 mit normalem mdadm

ydtxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdd1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

fqga 4 ephj 8 mcli 12 sunrkv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdf1:1024 /dev/sdq1:1024 /dev/sdg1:1024 /dev/sda1:1024 /dev/sde1:1024 /dev/sdp1:1024 /dev/sdh1:1024 /dev/sdj1:1024 /dev/sdm1:1024 /dev/sdc1:1024 /dev/sdl1:1024 /dev/sdi1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdk1:1024 /dev/sdv1:1024

19. Raid Recreate DATA und DATA2 2015-05-24

DATA2 mit normalem mdadm

yctxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdc1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

fqgm 4 ephj 8 dbli 12 sunrkv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdf1:1024 /dev/sdq1:1024 /dev/sdg1:1024 /dev/sdm1:1024 /dev/sde1:1024 /dev/sdp1:1024 /dev/sdh1:1024 /dev/sdj1:1024 /dev/sdd1:1024 /dev/sdb1:1024 /dev/sdl1:1024 /dev/sdi1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdk1:1024 /dev/sdv1:1024

20. Raid Recreate DATA und DATA2 2015-05-31

DATA2 mit normalem mdadm

ydtxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdd1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

eqfl 4 bpgi 8 mckh 12 sunrjv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sde1:1024 /dev/sdq1:1024 /dev/sdf1:1024 /dev/sdl1:1024 /dev/sdb1:1024 /dev/sdp1:1024 /dev/sdg1:1024 /dev/sdi1:1024 /dev/sdm1:1024 /dev/sdc1:1024 /dev/sdk1:1024 /dev/sdh1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdj1:1024 /dev/sdv1:1024

21. Raid Recreate DATA und DATA2 2015-06-06

DATA2 mit normalem mdadm

ydtxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdd1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

eqfl 4 bpgi 8 mckh 12 sunrjv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sde1:1024 /dev/sdq1:1024 /dev/sdf1:1024 /dev/sdl1:1024 /dev/sdb1:1024 /dev/sdp1:1024 /dev/sdg1:1024 /dev/sdi1:1024 /dev/sdm1:1024 /dev/sdc1:1024 /dev/sdk1:1024 /dev/sdh1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdj1:1024 /dev/sdv1:1024

22. Raid Recreate DATA und DATA2 2015-06-09

DATA2 mit normalem mdadm

yctxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdc1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

eqfl 4 dpgi 8 mbkh 12 sunrjv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sde1:1024 /dev/sdq1:1024 /dev/sdf1:1024 /dev/sdl1:1024 /dev/sdd1:1024 /dev/sdp1:1024 /dev/sdg1:1024 /dev/sdi1:1024 /dev/sdm1:1024 /dev/sdb1:1024 /dev/sdk1:1024 /dev/sdh1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdj1:1024 /dev/sdv1:1024

23. Raid Recreate DATA und DATA2 2015-06-11

DATA2 mit normalem mdadm

yctxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdc1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

eqfl 4 dpgi 8 mbkh 12 sunrjv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sde1:1024 /dev/sdq1:1024 /dev/sdf1:1024 /dev/sdl1:1024 /dev/sdd1:1024 /dev/sdp1:1024 /dev/sdg1:1024 /dev/sdi1:1024 /dev/sdm1:1024 /dev/sdb1:1024 /dev/sdk1:1024 /dev/sdh1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdj1:1024 /dev/sdv1:1024

24. Raid Recreate DATA und DATA2 2015-06-12

DATA2 mit normalem mdadm

yctxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdc1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

eqfl 4 dpgi 8 mbkh 12 sunrjv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sde1:1024 /dev/sdq1:1024 /dev/sdf1:1024 /dev/sdl1:1024 /dev/sdd1:1024 /dev/sdp1:1024 /dev/sdg1:1024 /dev/sdi1:1024 /dev/sdm1:1024 /dev/sdb1:1024 /dev/sdk1:1024 /dev/sdh1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdj1:1024 /dev/sdv1:1024

25. Raid Recreate DATA und DATA2 2015-06-20, irgendwie in letztet Zeit immer nur, wenn die Tür geöffnet/bewegt wurde

DATA2 mit normalem mdadm

ydtxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdd1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

eqfl 4 bpgi 8 mckh 12 sunrjv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sde1:1024 /dev/sdq1:1024 /dev/sdf1:1024 /dev/sdl1:1024 /dev/sdb1:1024 /dev/sdp1:1024 /dev/sdg1:1024 /dev/sdi1:1024 /dev/sdm1:1024 /dev/sdc1:1024 /dev/sdk1:1024 /dev/sdh1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdj1:1024 /dev/sdv1:1024

26. Raid Recreate DATA und DATA2 2015-06-27, irgendwie in letztet Zeit immer nur, wenn die Tür geöffnet/bewegt wurde

DATA2 mit normalem mdadm

ydtxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdd1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

cqfl 4 epgi 8 mbkh 12 sunrjv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdc1:1024 /dev/sdq1:1024 /dev/sdf1:1024 /dev/sdl1:1024 /dev/sde1:1024 /dev/sdp1:1024 /dev/sdg1:1024 /dev/sdi1:1024 /dev/sdm1:1024 /dev/sdb1:1024 /dev/sdk1:1024 /dev/sdh1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdj1:1024 /dev/sdv1:1024

27. Raid Recreate DATA und DATA2 2015-06-28

DATA2 mit normalem mdadm

ybtxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdb1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

dqel 4 cpfh 8 makg 12 sunrjv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdd1:1024 /dev/sdq1:1024 /dev/sde1:1024 /dev/sdl1:1024 /dev/sdc1:1024 /dev/sdp1:1024 /dev/sdf1:1024 /dev/sdh1:1024 /dev/sdm1:1024 /dev/sda1:1024 /dev/sdk1:1024 /dev/sdg1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdj1:1024 /dev/sdv1:1024

28. Raid Recreate DATA und DATA2 2015-07-14

DATA2 mit normalem mdadm

ydtxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdd1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

fqgm 4 ephj 8 bcli 12 sunrkv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdf1:1024 /dev/sdq1:1024 /dev/sdg1:1024 /dev/sdm1:1024 /dev/sde1:1024 /dev/sdp1:1024 /dev/sdh1:1024 /dev/sdj1:1024 /dev/sdb1:1024 /dev/sdc1:1024 /dev/sdl1:1024 /dev/sdi1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdk1:1024 /dev/sdv1:1024

29. Raid Recreate DATA und DATA2 2015-07-19

DATA2 mit normalem mdadm

xbswv
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdx1 /dev/sdb1 /dev/sds1 /dev/sdw1 /dev/sdv1 

DATA mdadm aus /builds/mdadm

dpek 4 cofh 8 lajg 12 rtmqiu
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdd1:1024 /dev/sdp1:1024 /dev/sde1:1024 /dev/sdk1:1024 /dev/sdc1:1024 /dev/sdo1:1024 /dev/sdf1:1024 /dev/sdh1:1024 /dev/sdl1:1024 /dev/sda1:1024 /dev/sdj1:1024 /dev/sdg1:1024 /dev/sdr1:1024 /dev/sdt1:1024 /dev/sdm1:1024 /dev/sdq1:1024 /dev/sdi1:1024 /dev/sdu1:1024

30. Raid Recreate DATA und DATA2 2015-07-26

DATA2 mit normalem mdadm

yctxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdc1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

eqfl 4 dpgi 8 mbkh 12 sunrjv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sde1:1024 /dev/sdq1:1024 /dev/sdf1:1024 /dev/sdl1:1024 /dev/sdd1:1024 /dev/sdp1:1024 /dev/sdg1:1024 /dev/sdi1:1024 /dev/sdm1:1024 /dev/sdb1:1024 /dev/sdk1:1024 /dev/sdh1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdj1:1024 /dev/sdv1:1024

31. Raid Recreate DATA und DATA2 2015-07-27

DATA2 mit normalem mdadm

ybtxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdb1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

dqek 4 cpfh 8 lajg 12 sunriv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdd1:1024 /dev/sdq1:1024 /dev/sde1:1024 /dev/sdk1:1024 /dev/sdc1:1024 /dev/sdp1:1024 /dev/sdf1:1024 /dev/sdh1:1024 /dev/sdl1:1024 /dev/sda1:1024 /dev/sdj1:1024 /dev/sdg1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdi1:1024 /dev/sdv1:1024

32. Raid Recreate DATA und DATA2 2015-08-09

DATA2 mit normalem mdadm

xbswv
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdx1 /dev/sdb1 /dev/sds1 /dev/sdw1 /dev/sdv1 

DATA mdadm aus /builds/mdadm

epfl 4 dogi 8 cakh 12 rtmqju
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sde1:1024 /dev/sdp1:1024 /dev/sdf1:1024 /dev/sdl1:1024 /dev/sdd1:1024 /dev/sdo1:1024 /dev/sdg1:1024 /dev/sdi1:1024 /dev/sdc1:1024 /dev/sda1:1024 /dev/sdk1:1024 /dev/sdh1:1024 /dev/sdr1:1024 /dev/sdt1:1024 /dev/sdm1:1024 /dev/sdq1:1024 /dev/sdj1:1024 /dev/sdu1:1024

33. Raid Recreate DATA und DATA2 2015-08-20

DATA2 mit normalem mdadm

ybtxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdb1 /dev/sdt1 /dev/sdx1 /dev/sdw1 

DATA mdadm aus /builds/mdadm

dqel 4 cpfh 8 makg 12 sunrjv
./mdadm --create --assume-clean /dev/md125 --chunk=64 --level=raid6 --layout=ls --raid-devices=18 /dev/sdd1:1024 /dev/sdq1:1024 /dev/sde1:1024 /dev/sdl1:1024 /dev/sdc1:1024 /dev/sdp1:1024 /dev/sdf1:1024 /dev/sdh1:1024 /dev/sdm1:1024 /dev/sda1:1024 /dev/sdk1:1024 /dev/sdg1:1024 /dev/sds1:1024 /dev/sdu1:1024 /dev/sdn1:1024 /dev/sdr1:1024 /dev/sdj1:1024 /dev/sdv1:1024

34. Raid Recreate NUR DATA2 2015-09-01 NACH NETZTEILWECHSEL

DATA2 mit normalem mdadm

ybtxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdb1 /dev/sdt1 /dev/sdx1 /dev/sdw1

35. Raid Recreate NUR DATA2 2015-09-01 NACH NETZTEILWECHSEL nach interupted DATA2 Raid6 Grow

Grow
mdadm --grow --raid-devices=6 /dev/md126 --backup-file=/home/gagi/mda126backup_20150911  

DATA2 mit normalem mdadm

ydtxw
 mdadm --create --assume-clean /dev/md126 --chunk=64 --level=raid6 --layout=ls --raid-devices=5 /dev/sdy1 /dev/sdd1 /dev/sdt1 /dev/sdx1 /dev/sdw1
Neu dazu kam dann sdo1

Versuch von --update=revert-reshape

--assemble --update=revert-reshape" can be used to undo a reshape
 that has just been started but isn't really wanted.  This is very
 new and while it passes basic tests it cannot be guaranteed.

mdadm --assemble --update=revert-reshape /dev/md126 /dev/sdy1 /dev/sdd1 /dev/sdt1 /dev/sdx1 /dev/sdw1 /dev/sdo1 --backup-file=/home/gagi/mda126backup_20150911

Spare-Group einrichten

aktuelle Config in mdadm.conf schreiben

mdadm -D -s >> /etc/mdadm.conf 

spare-group ergänzen

nano /etc/mdadm.conf

ganz unten spare-group=shared ergänzen

ARRAY /dev/md/126 metadata=1.2 spares=1 name=store2:126 UUID=xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx spare-group=shared

Raid-Baustatus

cat /proc/mdstat

automatisch jede Sekunde aktualisiert

watch -n 1 cat /proc/mdstat

Verschlüssel von Hand (ohne Yast2)

Verschlüsseln:

cryptsetup -v --key-size 256 luksFormat /dev/md125 

Mit speziellen Paramtern für Hardware-Verschlüsselung:

cryptsetup -v luksFormat --cipher aes-cbc-essiv:sha256 --key-size 256 /dev/md125

Öffnen:

cryptsetup luksOpen /dev/md125 cr_md125 

Filesystem drauf:

mkfs.xfs /dev/mapper/cr_md125 
store2:~ # mkfs.xfs /dev/mapper/cr_md125
meta-data=/dev/mapper/cr_md125   isize=256    agcount=36, agsize=268435424 blks
         =                       sectsz=512   attr=2, projid32bit=0
data     =                       bsize=4096   blocks=9523357168, imaxpct=5
         =                       sunit=16     swidth=208 blks
naming   =version 2              bsize=4096   ascii-ci=0
log      =internal log           bsize=4096   blocks=521728, version=2
         =                       sectsz=512   sunit=16 blks, lazy-count=1
realtime =none                   extsz=4096   blocks=0, rtextents=0


Status:

cryptsetup luksDump /dev/md125

Grown

Festplatte vorbereiten

Open gdisk with the first hard drive:
$ gdisk /dev/sda 

and type the following commands at the prompt:

   Add a new partition: n
   Select the default partition number: Enter
   Use the default for the first sector: Enter
   For sda1 and sda2 type the appropriate size in MB (i.e. +100MB and +2048M). For sda3 just hit Enter to select the remainder of the disk.
   Select Linux RAID as the partition type: fd00
   Write the table to disk and exit: w 

Bad Blocks

Screen-Umgebung starten

screen -S bb
badblocks -vs -o sdy-badblock-test /dev/sdy

verbose, show progress, output-file (log) badblocks sucht nur nach bad blocks, zerstört aber keine Daten.


Detach

Strg-a d = detach

Reatach

screen -r bb

ODER Wieder reingehen

screen -x bb

Device zum Raid hinzufügen

mdadm --add /dev/md125 /dev/sdc1

Raid reshapen mit zusätzlichem Device (dauert ca. 3 volle Tage)

mdadm --grow --raid-devices=18 /dev/md125 --backup-file=/home/gagi/mda125backup

um zu sehen, wer/was gerade Zugriff nimmt:

lsof /data 

Samba-Service beenden:

rcsmb stop 
systemctl stop smbd 

Unmounten:

umount /data 


XFS (Data)

XFS checken (ungemountet)

xfs_repair -n -o bhash=1024 /dev/mapper/cr_md125 

Cryptcontainer wachsen

cryptsetup --verbose resize cr_md125 

Mounten:

mount /dev/mapper/cr_md125 /data

XFS vergrößern

xfs_growfs /data 

XFS checken (ungemountet)

xfs_repair -n -o bhash=1024 /dev/mapper/cr_md125 


Read-Only Mounten:

mount -o ro /dev/mapper/cr_md125 /data

Samba-Freigabe

There's a bug in Samba in openSuse 11.4. Here's the workaround:

   go to Yast --> AppArmor --> Control Panel (on) --> Configure Profile Modes --> usr.sbin.smbd = complain
   go to Yast --> system --> runlevels --> smb=on + nmb=on
   reboot

Direkte Netwerkverbindung Store1 <-> Store 2.0

du kannst auch mal schauen was in /etc/udev/rules.d/70-persistent-net... (oder wie auch immer die date heißt) steht.

da wird die mac einer bestimmten netzwerkadresse (eth0, eth1, ...) zugewiesen.

die datei kannst du auch löschen oder verschieben - wird beim neustart neu angelegt.

da kommt machmal was durcheinander - bei 'nem kernelupdate oder bios-update.

GEHT ! unterschiedliche subnets (192.168.2.100 und 192.168.2.102)

Fast-Copy

1.) Empfänger (Store2.0)

cd <Zielverzeichnis> 
netcat -l -p 4323 | gunzip | cpio -i -d -m 

2.) Sender (Store)

cd <Quellverzeichnis>
find . -type f | cpio -p | gzip -1 | netcat 192.168.2.102 4323





1.) Empfänger (Store2.0)

socat tcp4-listen:4323 stdout | tar xvpf - /data/eBooks

2.) Sender (Store)

tar cvf - /data/eBooks | socat stdin tcp4:192.168.2.102:4323 


Test mit Fortschrittsanzeige bei bekannter Datengröße:

1.) Empfänger (Store2.0)

cd <Zielverzeichnis> 
socat tcp4-listen:4323 stdout | pv -s 93G | tar xvpf -

2.) Sender (Store)

cd <Quellverzeichnis> 
tar cvf - * | pv -s 93G | socat stdin tcp4:192.168.2.102:4323


dd if=/dev/sdl | bar -s 1.5T | dd of=/dev/sdw

FileBot Renamer Linux

filebot -rename -get-subtitles -non-strict /data/Downloads/Fertig/ --output /data/Downloads/Fertig/FileBot/ --format "{n}/Season {s}/{n}.{s00e00}.{t}" --db TheTVDB 
filebot -get-missing-subtitles -non-strict -r --lang en /data/Downloads/Fertig/FileBot/ 
filebot -script fn:replace --conflict override --def "e=.eng.srt" "r=.srt" /data/Downloads/Fertig/FileBot/

RemoteDesktop ArchLinux Client

z.B. auf Busfahrer: 
rdesktop -g 1440x900 -P -z  -x l -r sound:off -u gagi 192.168.1.149

Backplane-Rotation zur Fehlerdiagnose

Urzustand mit 24 Platten 2013-11-12

/dev/sdx -> 00 : 00000000000000 (1TB) 31��C
/dev/sdp -> 01 : WD-WCC070299387 (3TB WD) 31��C
/dev/sdq -> 03 : MJ1311YNG3SSLA (3TB) 33��C
/dev/sdr -> 05 : MJ1311YNG3NZ3A (3TB) 32��C
/dev/sds -> 07 : MJ1311YNG4J48A (3TB) 32��C
/dev/sdt -> 09 : MJ1311YNG3UUPA (3TB) 33��C
/dev/sdu -> 11 : MJ1311YNG3SAMA (3TB) 32��C
/dev/sdv -> 13 : MJ1311YNG3SU1A (3TB) 34��C
/dev/sdw -> 15 : MCE9215Q0AUYTW (3TB Toshiba neu) 31��C
/dev/sdh -> 16 : MJ0351YNGA02YA (3TB) nicht im Einsatz, bb-check 2013-08-28 37��C
/dev/sdi -> 18 : MJ1311YNG3Y4SA (3TB) 40��C
/dev/sdj -> 20 : WD-WCAWZ1881335 (3TB WD) hot spare 38��C
/dev/sdk -> 22 : WD-WCAWZ2279670 (3TB WD) 41��C
/dev/sdl -> 24 : MJ1311YNG25Z6A (3TB) 39��C
/dev/sdm -> 26 : MJ1311YNG3RM5A (3TB) 39��C
/dev/sdn -> 28 : MJ1311YNG3NT5A (3TB) 40��C
/dev/sdo -> 30 : MCM9215Q0B9LSY (3TB Toshiba neu) 38��C
/dev/sda -> 31 : 234BGY0GS (3TB Toshiba neu) 40��C
/dev/sdb -> 33 : MJ1311YNG3WZVA (3TB) 43��C
/dev/sdc -> 35 : MJ1311YNG3SYKA (3TB) 42��C
/dev/sdd -> 37 : WD-WCC070198169 (3TB WD) 41��C
/dev/sde -> 39 : MJ1311YNG3RZTA (3TB) 39��C
/dev/sdf -> 41 : MJ1311YNG3LTRA (3TB) 39��C
/dev/sdg -> 43 : MJ1311YNG38VGA (3TB) 39��C
Insgesamt 24 Platten gefunden.

Crashes

ArchLinux 2011-09-09

Sep  9 18:20:04 localhost kernel: [156439.479947] program smartctl is using a deprecated SCSI ioctl, please convert it to SG_IO
Sep  9 18:20:04 localhost kernel: [156439.480035] program smartctl is using a deprecated SCSI ioctl, please convert it to SG_IO
Sep  9 18:20:04 localhost kernel: [156439.486612] program smartctl is using a deprecated SCSI ioctl, please convert it to SG_IO
Sep  9 18:20:04 localhost kernel: [156439.503656] program smartctl is using a deprecated SCSI ioctl, please convert it to SG_IO
Sep  9 18:20:04 localhost kernel: [156439.504562] program smartctl is using a deprecated SCSI ioctl, please convert it to SG_IO
Sep  9 18:34:11 localhost -- MARK --
Sep  9 18:42:42 localhost kernel: [157797.911330] r8169: eth0: link up
Sep  9 18:54:11 localhost -- MARK --
Sep  9 19:14:11 localhost -- MARK --
Sep  9 19:34:11 localhost -- MARK --
Sep  9 19:54:11 localhost -- MARK --
Sep  9 20:14:11 localhost -- MARK --
Sep  9 20:27:32 localhost kernel: [164086.971566] r8169: eth0: link up
Sep  9 20:27:42 localhost kernel: [164097.580071] r8169: eth0: link up
Sep  9 20:27:50 localhost kernel: [164105.391755] r8169: eth0: link up
Sep  9 20:27:51 localhost kernel: [164106.272019] r8169: eth0: link up
Sep  9 20:28:12 localhost kernel: [164127.150062] r8169: eth0: link up
Sep  9 20:28:22 localhost kernel: [164137.941304] r8169: eth0: link up
Sep  9 20:28:33 localhost kernel: [164148.890097] r8169: eth0: link up
Sep  9 20:28:38 localhost kernel: [164153.080536] r8169: eth0: link up
Sep  9 20:28:58 localhost kernel: [164173.790064] r8169: eth0: link up
Sep  9 20:42:19 localhost kernel: [    0.000000] Initializing cgroup subsys cpuset
Sep  9 20:42:19 localhost kernel: [    0.000000] Initializing cgroup subsys cpu
Sep  9 20:42:19 localhost kernel: [    0.000000] Linux version 2.6.32-lts (tobias@T-POWA-LX) (gcc version 4.6.1 20110819    (prerelease) (GCC) ) #1 SMP Tue Aug 30 08:59:44 CEST 2011
Sep  9 20:42:19 localhost kernel: [    0.000000] Command line: root=/dev/disk/by-uuid/ba47ea9a-c24c-4dc6-a9a2-ca3b442bdbfc ro vga=0x31B
Sep  9 20:42:19 localhost kernel: [    0.000000] KERNEL supported cpus:
Sep  9 20:42:19 localhost kernel: [    0.000000]   Intel GenuineIntel
Sep  9 20:42:19 localhost kernel: [    0.000000]   AMD AuthenticAMD
Sep  9 20:42:19 localhost kernel: [    0.000000]   Centaur CentaurHauls

OpenSuse 2011-09-26

Sep 26 23:15:59 store2 su: (to nobody) root on none
Sep 26 23:17:17  su: last message repeated 2 times
Sep 26 23:25:23 store2 smartd[4617]: Device: /dev/sdc [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 166 to 162
Sep 26 23:25:26 store2 smartd[4617]: Device: /dev/sde [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 171 to 166
Sep 26 23:25:29 store2 smartd[4617]: Device: /dev/sdg [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 171 to 166
Sep 26 23:25:36 store2 smartd[4617]: Device: /dev/sdk [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 171 to 166
Sep 26 23:25:37 store2 smartd[4617]: Device: /dev/sdl [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 181 to 187
Sep 26 23:55:22 store2 smartd[4617]: Device: /dev/sdb [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 100 to 96
Sep 26 23:55:23 store2 smartd[4617]: Device: /dev/sdc [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 162 to 166
Sep 26 23:55:26 store2 smartd[4617]: Device: /dev/sde [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 100 to 99
Sep 26 23:55:26 store2 smartd[4617]: Device: /dev/sde [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 166 to 171
Sep 26 23:55:29 store2 smartd[4617]: Device: /dev/sdg [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 166 to 171
Sep 26 23:55:32 store2 smartd[4617]: Device: /dev/sdi [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 100 to 98
Sep 27 00:55:26 store2 kernel: imklog 5.6.5, log source = /proc/kmsg started.

OpenSuse2011-09-27

Sep 27 16:35:17 store2 smbd[29588]: [2011/09/27 16:35:17.391212,  0] param/loadparm.c:8445(check_usershare_stat)
Sep 27 16:35:17 store2 smbd[29588]:   check_usershare_stat: file /var/lib/samba/usershares/ owned by uid 0 is not a regular file
Sep 27 16:44:06 store2 smbd[29163]: [2011/09/27 16:44:06.795153,  0] lib/util_sock.c:474(read_fd_with_timeout)
Sep 27 16:44:06 store2 smbd[29163]: [2011/09/27 16:44:06.795341,  0] lib/util_sock.c:1441(get_peer_addr_internal)
Sep 27 16:44:06 store2 smbd[29597]: [2011/09/27 16:44:06.795323,  0] lib/util_sock.c:474(read_fd_with_timeout)
Sep 27 16:44:06 store2 smbd[29163]:   getpeername failed. Error was Der Socket ist nicht verbunden
Sep 27 16:44:06 store2 smbd[29592]: [2011/09/27 16:44:06.795368,  0] lib/util_sock.c:474(read_fd_with_timeout)
Sep 27 16:44:06 store2 smbd[29163]:   read_fd_with_timeout: client 0.0.0.0 read error = Die Verbindung wurde vom Kommunikationspartner zurückgesetzt.
Sep 27 16:44:06 store2 smbd[29597]: [2011/09/27 16:44:06.795422,  0] lib/util_sock.c:1441(get_peer_addr_internal)
Sep 27 16:44:06 store2 smbd[29597]:   getpeername failed. Error was Der Socket ist nicht verbunden
Sep 27 16:44:06 store2 smbd[29597]:   read_fd_with_timeout: client 0.0.0.0 read error = Die Verbindung wurde vom Kommunikationspartner zurückgesetzt.
Sep 27 16:44:06 store2 smbd[29592]: [2011/09/27 16:44:06.795468,  0] lib/util_sock.c:1441(get_peer_addr_internal)
Sep 27 16:44:06 store2 smbd[29592]:   getpeername failed. Error was Der Socket ist nicht verbunden
Sep 27 16:44:06 store2 smbd[29592]:   read_fd_with_timeout: client 0.0.0.0 read error = Die Verbindung wurde vom Kommunikationspartner zurückgesetzt.
Sep 27 16:45:42 store2 smbd[29585]: [2011/09/27 16:45:42.499038,  0] lib/util_sock.c:474(read_fd_with_timeout)
Sep 27 16:45:42 store2 smbd[29593]: [2011/09/27 16:45:42.499082,  0] lib/util_sock.c:474(read_fd_with_timeout)
Sep 27 16:45:42 store2 smbd[29593]: [2011/09/27 16:45:42.499174,  0] lib/util_sock.c:1441(get_peer_addr_internal)
Sep 27 16:45:42 store2 smbd[29585]: [2011/09/27 16:45:42.499174,  0] lib/util_sock.c:1441(get_peer_addr_internal)
Sep 27 16:45:42 store2 smbd[29593]:   getpeername failed. Error was Der Socket ist nicht verbunden
Sep 27 16:45:42 store2 smbd[29585]:   getpeername failed. Error was Der Socket ist nicht verbunden
Sep 27 16:45:42 store2 smbd[29593]:   read_fd_with_timeout: client 0.0.0.0 read error = Die Verbindung wurde vom Kommunikationspartner zurückgesetzt.
Sep 27 16:45:42 store2 smbd[29585]:   read_fd_with_timeout: client 0.0.0.0 read error = Die Verbindung wurde vom Kommunikationspartner zurückgesetzt.
Sep 27 19:35:14 store2 kernel: imklog 5.6.5, log source = /proc/kmsg started.

OpenSuse 2011-09-29

während kräftigem Copyjob von Store

Sep 29 23:16:19  su: last message repeated 2 times
Sep 29 23:28:41 store2 smartd[4624]: Device: /dev/sdb [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 157 to 162
Sep 29 23:28:44 store2 smartd[4624]: Device: /dev/sdd [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 153 to 157
Sep 29 23:28:49 store2 smartd[4624]: Device: /dev/sdh [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 100 to 99
Sep 29 23:28:53 store2 smartd[4624]: Device: /dev/sdk [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 162 to 157
Sep 29 23:28:57 store2 smartd[4624]: Device: /dev/sdo [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 176 to 181
Sep 29 23:58:44 store2 smartd[4624]: Device: /dev/sdd [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 100 to 99
Sep 29 23:58:49 store2 smartd[4624]: Device: /dev/sdh [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 99 to 100
Sep 29 23:58:53 store2 smartd[4624]: Device: /dev/sdk [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 157 to 162
Sep 29 23:58:57 store2 smartd[4624]: Device: /dev/sdn [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 187 to 193
Sep 29 23:58:58 store2 smartd[4624]: Device: /dev/sdo [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 181 to 176
Sep 29 23:59:02 store2 smartd[4624]: Device: /dev/sdq [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 176 to 181
Sep 30 00:28:41 store2 smartd[4624]: Device: /dev/sda [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 187 to 193
Sep 30 00:28:43 store2 smartd[4624]: Device: /dev/sdd [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 99 to 100
Sep 30 00:28:49 store2 smartd[4624]: Device: /dev/sdh [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 100 to 99
Sep 30 00:28:58 store2 smartd[4624]: Device: /dev/sdo [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 176 to 181
Sep 30 00:58:47 store2 smartd[4624]: Device: /dev/sdf [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 100 to 99
Sep 30 00:58:49 store2 smartd[4624]: Device: /dev/sdh [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 99 to 100
Sep 30 00:58:59 store2 smartd[4624]: Device: /dev/sdp [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 100 to 99
Sep 30 01:28:47 store2 smartd[4624]: Device: /dev/sdf [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 99 to 100
Sep 30 01:28:47 store2 smartd[4624]: Device: /dev/sdf [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 157 to 162
Sep 30 01:28:50 store2 smartd[4624]: Device: /dev/sdi [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 100 to 99
Sep 30 01:58:47 store2 smartd[4624]: Device: /dev/sdf [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 162 to 157
Sep 30 01:59:00 store2 smartd[4624]: Device: /dev/sdp [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 99 to 100
Sep 30 02:28:45 store2 smartd[4624]: Device: /dev/sdd [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 100 to 99
Sep 30 02:28:46 store2 smartd[4624]: Device: /dev/sde [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 100 to 99
Sep 30 02:28:46 store2 smartd[4624]: Device: /dev/sde [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 157 to 162
Sep 30 02:28:48 store2 smartd[4624]: Device: /dev/sdf [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 100 to 99
Sep 30 02:28:52 store2 smartd[4624]: Device: /dev/sdi [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 99 to 100
Sep 30 02:58:45 store2 smartd[4624]: Device: /dev/sdd [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 99 to 100
Sep 30 02:58:46 store2 smartd[4624]: Device: /dev/sde [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 99 to 100
Sep 30 02:58:46 store2 smartd[4624]: Device: /dev/sde [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 162 to 157 
Sep 30 02:58:47 store2 smartd[4624]: Device: /dev/sdf [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 99 to 100
Sep 30 02:58:49 store2 smartd[4624]: Device: /dev/sdh [SAT], SMART Prefailure Attribute: 1 Raw_Read_Error_Rate changed from 100 to 99
Sep 30 09:39:22 store2 kernel: imklog 5.6.5, log source = /proc/kmsg started.

What you are seeing are the Normalized Attribute values changing.

For example when the Raw_Read_Error_Rate changed from 99 to 100, the increase in Normalized value from 99 to 100 means that the disk now thinks it is a bit LESS likely to fail than before, because this Normalized value is moving further above the (low) Threshold value.

ArchLinux 2011-10-17

Oct 17 21:21:35 localhost smartd[1941]: Device: /dev/sdc [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 181 to 176
Oct 17 21:21:37 localhost smartd[1941]: Device: /dev/sde [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 181 to 176
Oct 17 21:21:45 localhost smartd[1941]: Device: /dev/sdm [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 200 to 206
Oct 17 21:30:03 localhost -- MARK --
Oct 17 21:50:03 localhost -- MARK --
Oct 17 21:51:37 localhost smartd[1941]: Device: /dev/sde [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 176 to 181
Oct 17 21:51:41 localhost smartd[1941]: Device: /dev/sdi [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 181 to 187
Oct 17 22:10:03 localhost -- MARK --
Oct 17 22:21:34 localhost smartd[1941]: Device: /dev/sdc [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 176 to 181
Oct 17 22:21:47 localhost smartd[1941]: Device: /dev/sdo [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 206 to 214
Oct 17 22:21:49 localhost smartd[1941]: Device: /dev/sdq [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 206 to 214
Oct 17 22:30:03 localhost -- MARK --
Oct 17 22:50:03 localhost -- MARK --
Oct 17 23:11:18 localhost kernel: [    0.000000] Initializing cgroup subsys cpuset
Oct 17 23:11:18 localhost kernel: [    0.000000] Initializing cgroup subsys cpu

ArchLinux 2011-11-06

Nov  6 12:39:05 localhost -- MARK --
Nov  6 12:42:18 localhost smartd[1927]: Device: /dev/sdi [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 193 to 200
Nov  6 12:42:20 localhost smartd[1927]: Device: /dev/sdj [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 193 to 187
Nov  6 12:42:24 localhost smartd[1927]: Device: /dev/sdn [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 222 to 214
Nov  6 12:42:25 localhost smartd[1927]: Device: /dev/sdo [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 230 to 222
Nov  6 12:42:26 localhost smartd[1927]: Device: /dev/sdp [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 214 to 222
Nov  6 12:59:05 localhost -- MARK --
Nov  6 14:29:21 localhost kernel: [    0.000000] Initializing cgroup subsys cpuset
Nov  6 14:29:21 localhost kernel: [    0.000000] Initializing cgroup subsys cpu
Nov  6 14:29:21 localhost kernel: [    0.000000] Linux version 3.0-ARCH (tobias@T-POWA-LX) (gcc version 4.6.1 20110819 (prerelease) (GCC) ) #1 SMP PREEMPT Wed Oct$
Nov  6 14:29:21 localhost kernel: [    0.000000] Command line: root=/dev/disk/by-id/ata-Hitachi_HCS5C1016CLA382_JC0150HT0J7TPC-part3 ro
Nov  6 14:29:21 localhost kernel: [    0.000000] BIOS-provided physical RAM map: 

EINFACH SO!

ArchLinux 2011-11-21

Dabei war vorher ein Systemupdate gelaufen (inklusive neuem Kernel), aber noch nicht rebootet.

Nov 21 09:30:27 localhost smartd[2208]: Device: /dev/sdj [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 200 to 206
Nov 21 09:30:30 localhost smartd[2208]: Device: /dev/sdl [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 240 to 250
Nov 21 09:30:31 localhost smartd[2208]: Device: /dev/sdm [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 250 to 240
Nov 21 09:30:35 localhost smartd[2208]: Device: /dev/sdp [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 230 to 240
Nov 21 09:43:32 localhost kernel: [1280595.864622] ------------[ cut here ]------------
Nov 21 09:43:32 localhost kernel: [1280595.864636] WARNING: at drivers/gpu/drm/i915/i915_irq.c:649 ironlake_irq_handler+0x1102/0x1110 [i915]()
Nov 21 09:43:32 localhost kernel: [1280595.864638] Hardware name: H61M-S2V-B3
Nov 21 09:43:32 localhost kernel: [1280595.864639] Missed a PM interrupt
Nov 21 09:43:32 localhost kernel: [1280595.864640] Modules linked in: xfs sha256_generic dm_crypt dm_mod raid456 async_raid6_recov async_pq raid6_pq async_xor xor async_memcpy async_tx md_mod coretemp nfsd exportfs nfs lockd fscache  auth_rpcgss nfs_acl sunrpc ipv6 ext2 sr_mod cdrom snd_hda_codec_realtek usb_storage uas sg evdev snd_hda_intel snd_hda_codec iTCO_wdt snd_hwdep snd_pcm snd_timer i915 snd drm_kms_helper drm pcspkr i2c_algo_bit r8169 ppdev i2c_i801 shp chp parport_pc intel_agp i2c_core pci_hotplug parport intel_gtt mei(C) soundcore snd_page_alloc processor button mii iTCO_vendor_support video aesni_intel cryptd aes_x86_64 aes_generic ext4 mbcache jbd2 crc16 usbhid hid sd_mod sata_sil24 ahci libahci libata scsi_mod ehci_hcd usbcore
Nov 21 09:43:32 localhost kernel: [1280595.864674] Pid: 0, comm: swapper Tainted: G         C  3.0-ARCH #1
Nov 21 09:43:32 localhost kernel: [1280595.864675] Call Trace:
Nov 21 09:43:32 localhost kernel: [1280595.864676]  <IRQ>  [<ffffffff8105c76f>] warn_slowpath_common+0x7f/0xc0
Nov 21 09:43:32 localhost kernel: [1280595.864684]  [<ffffffff8105c866>] warn_slowpath_fmt+0x46/0x50
Nov 21 09:43:32 localhost kernel: [1280595.864688]  [<ffffffff81078f7d>] ? queue_work+0x5d/0x70
Nov 21 09:43:32 localhost kernel: [1280595.864693]  [<ffffffffa0235a22>] ironlake_irq_handler+0x1102/0x1110 [i915]
Nov 21 09:43:32 localhost kernel: [1280595.864696]  [<ffffffff812a4bc5>] ? dma_issue_pending_all+0x95/0xa0
Nov 21 09:43:32 localhost kernel: [1280595.864699]  [<ffffffff81333db1>] ? net_rx_action+0x131/0x300
Nov 21 09:43:32 localhost kernel: [1280595.864702]  [<ffffffff810bf835>] handle_irq_event_percpu+0x75/0x2a0
Nov 21 09:43:32 localhost kernel: [1280595.864705]  [<ffffffff810bfaa5>] handle_irq_event+0x45/0x70
Nov 21 09:43:32 localhost kernel: [1280595.864707]  [<ffffffff810c21af>] handle_edge_irq+0x6f/0x120
Nov 21 09:43:32 localhost kernel: [1280595.864710]  [<ffffffff8100d9f2>] handle_irq+0x22/0x40
Nov 21 09:43:32 localhost kernel: [1280595.864712]  [<ffffffff813f66aa>] do_IRQ+0x5a/0xe0
Nov 21 09:43:32 localhost kernel: [1280595.864715]  [<ffffffff813f4393>] common_interrupt+0x13/0x13
Nov 21 09:43:32 localhost kernel: [1280595.864716]  <EOI>  [<ffffffff81273cdb>] ? intel_idle+0xcb/0x120
Nov 21 09:43:32 localhost kernel: [1280595.864720]  [<ffffffff81273cbd>] ? intel_idle+0xad/0x120
Nov 21 09:43:32 localhost kernel: [1280595.864723]  [<ffffffff81313d9d>] cpuidle_idle_call+0x9d/0x350
Nov 21 09:43:32 localhost kernel: [1280595.864726]  [<ffffffff8100a21a>] cpu_idle+0xba/0x100
Nov 21 09:43:32 localhost kernel: [1280595.864729]  [<ffffffff813d1eb2>] rest_init+0x96/0xa4
Nov 21 09:43:32 localhost kernel: [1280595.864731]  [<ffffffff81748c23>] start_kernel+0x3de/0x3eb
Nov 21 09:43:32 localhost kernel: [1280595.864733]  [<ffffffff81748347>] x86_64_start_reservations+0x132/0x136
Nov 21 09:43:32 localhost kernel: [1280595.864735]  [<ffffffff81748140>] ? early_idt_handlers+0x140/0x140
Nov 21 09:43:32 localhost kernel: [1280595.864737]  [<ffffffff8174844d>] x86_64_start_kernel+0x102/0x111
Nov 21 09:43:32 localhost kernel: [1280595.864738] ---[ end trace 01037f4ec3ec4ee5 ]---
Nov 21 10:00:16 localhost smartd[2208]: Device: /dev/sda [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 206 to 200
Nov 21 10:00:18 localhost smartd[2208]: Device: /dev/sdc [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 200 to 193
Nov 21 10:00:19 localhost smartd[2208]: Device: /dev/sdd [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 206 to 200
Nov 21 10:00:23 localhost smartd[2208]: Device: /dev/sdg [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 206 to 200
Nov 21 10:00:29 localhost smartd[2208]: Device: /dev/sdl [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 250 to 240
Nov 21 10:00:30 localhost smartd[2208]: Device: /dev/sdm [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 240 to 250
Nov 21 10:00:33 localhost smartd[2208]: Device: /dev/sdp [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 240 to 230
Nov 21 10:00:34 localhost smartd[2208]: Device: /dev/sdq [SAT], SMART Usage Attribute: 194 Temperature_Celsius changed from 240 to 250
Nov 21 11:52:01 localhost kernel: [    0.000000] Initializing cgroup subsys cpuset
Nov 21 11:52:01 localhost kernel: [    0.000000] Initializing cgroup subsys cpu