sas2iru.go 10 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407
  1. // Copyright 2019 Yunion
  2. //
  3. // Licensed under the Apache License, Version 2.0 (the "License");
  4. // you may not use this file except in compliance with the License.
  5. // You may obtain a copy of the License at
  6. //
  7. // http://www.apache.org/licenses/LICENSE-2.0
  8. //
  9. // Unless required by applicable law or agreed to in writing, software
  10. // distributed under the License is distributed on an "AS IS" BASIS,
  11. // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  12. // See the License for the specific language governing permissions and
  13. // limitations under the License.
  14. package sas2iru
  15. import (
  16. "fmt"
  17. "regexp"
  18. "strconv"
  19. "strings"
  20. "yunion.io/x/pkg/tristate"
  21. "yunion.io/x/pkg/util/stringutils"
  22. api "yunion.io/x/onecloud/pkg/apis/compute"
  23. "yunion.io/x/onecloud/pkg/baremetal/utils/raid"
  24. "yunion.io/x/onecloud/pkg/compute/baremetal"
  25. )
  26. type Mpt2SASRaidPhyDev struct {
  27. *raid.RaidBasePhyDev
  28. enclosure int
  29. slot int
  30. sector int
  31. block int
  32. }
  33. func newMpt2SASRaidPhyDev(adapter int) *Mpt2SASRaidPhyDev {
  34. b := raid.NewRaidBasePhyDev(baremetal.DISK_DRIVER_MPT2SAS)
  35. b.Adapter = adapter
  36. return &Mpt2SASRaidPhyDev{
  37. RaidBasePhyDev: b,
  38. slot: -1,
  39. enclosure: -1,
  40. block: -1,
  41. sector: -1,
  42. }
  43. }
  44. func (dev *Mpt2SASRaidPhyDev) parseLine(line string) bool {
  45. key, val := stringutils.SplitKeyValue(line)
  46. if key == "" {
  47. return false
  48. }
  49. switch key {
  50. case "Drive Type":
  51. if strings.HasSuffix(val, "_HDD") {
  52. dev.Rotate = tristate.True
  53. } else {
  54. dev.Rotate = tristate.False
  55. }
  56. case "Enclosure #":
  57. dev.enclosure, _ = strconv.Atoi(val)
  58. case "Slot #":
  59. dev.slot, _ = strconv.Atoi(val)
  60. case "Size (in MB)/(in sectors)":
  61. dat := strings.Split(val, "/")
  62. sz, _ := strconv.Atoi(dat[0])
  63. dev.Size = int64(sz)
  64. dev.sector, _ = strconv.Atoi(dat[1])
  65. dev.block = int(dev.Size * 1024 * 1024 / 7814037167)
  66. if dev.block > 4000 {
  67. dev.block = 4096
  68. } else {
  69. dev.block = 512
  70. }
  71. dev.Size = int64(dev.block * dev.sector / 1024 / 1024)
  72. case "Manufacturer", "Model Number", "Firmware Revision", "Serial No":
  73. if dev.Model == "" {
  74. dev.Model = val
  75. } else {
  76. dev.Model = fmt.Sprintf("%s %s", dev.Model, val)
  77. }
  78. case "State":
  79. if strings.Contains(strings.ToLower(val), "ready") {
  80. dev.Status = "online"
  81. } else {
  82. dev.Status = strings.ToLower(val)
  83. }
  84. default:
  85. return false
  86. }
  87. return true
  88. }
  89. func (dev *Mpt2SASRaidPhyDev) isComplete() bool {
  90. if !dev.RaidBasePhyDev.IsComplete() {
  91. return false
  92. }
  93. if dev.Size < 0 {
  94. return false
  95. }
  96. if dev.slot < 0 {
  97. return false
  98. }
  99. if dev.sector < 0 {
  100. return false
  101. }
  102. if dev.block < 0 {
  103. return false
  104. }
  105. return true
  106. }
  107. func (dev *Mpt2SASRaidPhyDev) ToBaremetalStorage(idx int) *baremetal.BaremetalStorage {
  108. s := dev.RaidBasePhyDev.ToBaremetalStorage(idx)
  109. s.Index = int64(idx)
  110. s.Slot = dev.slot
  111. s.Enclosure = dev.enclosure
  112. s.Block = int64(dev.block)
  113. s.Sector = int64(dev.sector)
  114. return s
  115. }
  116. func GetSpecString(dev *baremetal.BaremetalStorage) string {
  117. if dev.Enclosure < 0 {
  118. return fmt.Sprintf(":%d", dev.Slot)
  119. }
  120. return fmt.Sprintf("%d:%d", dev.Enclosure, dev.Slot)
  121. }
  122. type Mpt2SASRaidAdaptor struct {
  123. index int
  124. raid *Mpt2SASRaid
  125. devs []*Mpt2SASRaidPhyDev
  126. }
  127. func newMpt2SASRaidAdaptor(index int, raid *Mpt2SASRaid) *Mpt2SASRaidAdaptor {
  128. return &Mpt2SASRaidAdaptor{
  129. index: index,
  130. raid: raid,
  131. devs: make([]*Mpt2SASRaidPhyDev, 0),
  132. }
  133. }
  134. func (adapter *Mpt2SASRaidAdaptor) GetIndex() int {
  135. return adapter.index
  136. }
  137. func (adapter *Mpt2SASRaidAdaptor) ParsePhyDevs() error {
  138. cmd := adapter.raid.GetCommand(fmt.Sprintf("%d", adapter.index), "DISPLAY")
  139. ret, err := adapter.raid.term.Run(cmd)
  140. if err != nil {
  141. return fmt.Errorf("get physical device: %v", err)
  142. }
  143. for _, l := range ret {
  144. if strings.Contains(l, "RAID Support") && strings.ToLower(strings.TrimSpace(l[strings.Index(l, ":")+1:])) == "no" {
  145. return fmt.Errorf("No raid support")
  146. }
  147. }
  148. return adapter.parsePhyDevs(ret)
  149. }
  150. func (adapter *Mpt2SASRaidAdaptor) parsePhyDevs(lines []string) error {
  151. dev := newMpt2SASRaidPhyDev(adapter.index)
  152. for _, l := range lines {
  153. if dev.parseLine(l) && dev.isComplete() {
  154. adapter.devs = append(adapter.devs, dev)
  155. dev = newMpt2SASRaidPhyDev(adapter.index)
  156. }
  157. }
  158. return nil
  159. }
  160. func (adapter *Mpt2SASRaidAdaptor) GetDevices() []*baremetal.BaremetalStorage {
  161. ret := []*baremetal.BaremetalStorage{}
  162. for idx, dev := range adapter.devs {
  163. ret = append(ret, dev.ToBaremetalStorage(idx))
  164. }
  165. return ret
  166. }
  167. func (adapter *Mpt2SASRaidAdaptor) GetLogicVolumes() ([]*raid.RaidLogicalVolume, error) {
  168. cmd := adapter.raid.GetCommand(fmt.Sprintf("%d", adapter.index), "DISPLAY")
  169. ret, err := adapter.raid.term.Run(cmd)
  170. if err != nil {
  171. return nil, fmt.Errorf("GetLogicVolumes error: %v", err)
  172. }
  173. return adapter.parseLogicVolumes(ret)
  174. }
  175. func (adapter *Mpt2SASRaidAdaptor) parseLogicVolumes(lines []string) ([]*raid.RaidLogicalVolume, error) {
  176. lvIdx := []*raid.RaidLogicalVolume{}
  177. usedDevs := []*raid.RaidLogicalVolume{}
  178. for _, line := range lines {
  179. key, val := stringutils.SplitKeyValue(line)
  180. if key != "" && key == "Volume ID" {
  181. idx, _ := strconv.Atoi(val)
  182. lvIdx = append(lvIdx, &raid.RaidLogicalVolume{
  183. Index: idx,
  184. Adapter: adapter.index,
  185. })
  186. } else if regexp.MustCompile(`PHY\[\d+\] Enclosure#/Slot#`).MatchString(key) {
  187. idx, _ := strconv.Atoi(val)
  188. usedDevs = append(usedDevs, &raid.RaidLogicalVolume{
  189. Index: idx,
  190. Adapter: adapter.index,
  191. })
  192. }
  193. }
  194. if len(adapter.devs) < len(usedDevs) {
  195. return nil, fmt.Errorf("adapter current dev %d < usedDevs %d", len(adapter.devs), len(usedDevs))
  196. }
  197. for i := 0; i < len(adapter.devs)-len(usedDevs); i++ {
  198. lvIdx = append(lvIdx, &raid.RaidLogicalVolume{
  199. Index: -1,
  200. Adapter: adapter.index,
  201. })
  202. }
  203. return lvIdx, nil
  204. }
  205. func (adapter *Mpt2SASRaidAdaptor) rescanLV() error {
  206. var cmd string
  207. if adapter.raid.utility == "/opt/lsi/sas2ircu" {
  208. cmd = "/opt/lsi/rescan.sh mpt2sas"
  209. } else if adapter.raid.utility == "/opt/lsi/sas3ircu" {
  210. cmd = "/opt/lsi/rescan.sh mpt3sas"
  211. } else {
  212. return fmt.Errorf("Unsupport raid utility: %v", adapter.raid.utility)
  213. }
  214. _, err := adapter.raid.term.Run(cmd)
  215. return err
  216. }
  217. func (adapter *Mpt2SASRaidAdaptor) setBootIR() error {
  218. lvs, err := adapter.GetLogicVolumes()
  219. if err != nil {
  220. return err
  221. }
  222. if len(lvs) > 0 && lvs[0].Index > 0 {
  223. args := []string{fmt.Sprintf("%d", adapter.index), "BOOTIR", fmt.Sprintf("%d", lvs[0].Index)}
  224. cmd := adapter.raid.GetCommand(args...)
  225. _, err := adapter.raid.term.Run(cmd)
  226. return err
  227. }
  228. return nil
  229. }
  230. func (adapter *Mpt2SASRaidAdaptor) PreBuildRaid(confs []*api.BaremetalDiskConfig) error {
  231. return nil
  232. }
  233. func (adapter *Mpt2SASRaidAdaptor) PostBuildRaid() error {
  234. return nil
  235. }
  236. func (adapter *Mpt2SASRaidAdaptor) buildRaid(level string, devs []*baremetal.BaremetalStorage, conf *api.BaremetalDiskConfig) error {
  237. if len(conf.Size) > 1 {
  238. return fmt.Errorf("Subdivide sub-size not supported")
  239. }
  240. args := []string{fmt.Sprintf("%d", adapter.index), "CREATE", level, "MAX"}
  241. labels := []string{}
  242. for _, dev := range devs {
  243. labels = append(labels, GetSpecString(dev))
  244. }
  245. args = append(args, labels...)
  246. args = append(args, "noprompt")
  247. _, err := adapter.raid.term.Run(adapter.raid.GetCommand(args...))
  248. if err != nil {
  249. return fmt.Errorf("Build raid error: %v", err)
  250. }
  251. if err := adapter.setBootIR(); err != nil {
  252. return fmt.Errorf("setBootIR: %v", err)
  253. }
  254. if err := adapter.rescanLV(); err != nil {
  255. return fmt.Errorf("rescanLV: %v", err)
  256. }
  257. return nil
  258. }
  259. func (adapter *Mpt2SASRaidAdaptor) BuildRaid0(devs []*baremetal.BaremetalStorage, conf *api.BaremetalDiskConfig) error {
  260. return adapter.buildRaid("RAID0", devs, conf)
  261. }
  262. func (adapter *Mpt2SASRaidAdaptor) BuildRaid1(devs []*baremetal.BaremetalStorage, conf *api.BaremetalDiskConfig) error {
  263. return adapter.buildRaid("RAID1", devs, conf)
  264. }
  265. func (adapter *Mpt2SASRaidAdaptor) BuildRaid5(devs []*baremetal.BaremetalStorage, conf *api.BaremetalDiskConfig) error {
  266. return fmt.Errorf("Not impl")
  267. }
  268. func (adapter *Mpt2SASRaidAdaptor) BuildRaid10(devs []*baremetal.BaremetalStorage, conf *api.BaremetalDiskConfig) error {
  269. if len(devs) > 10 {
  270. return fmt.Errorf("RAID10 supports no more than 10 disks")
  271. }
  272. return adapter.buildRaid("RAID10", devs, conf)
  273. }
  274. func (adapter *Mpt2SASRaidAdaptor) BuildNoneRaid(devs []*baremetal.BaremetalStorage) error {
  275. // TODO: not impl
  276. // return fmt.Errorf("Not impl")
  277. return nil
  278. }
  279. func (adapter *Mpt2SASRaidAdaptor) RemoveLogicVolumes() error {
  280. cmd := adapter.raid.GetCommand(fmt.Sprintf("%d", adapter.index), "DELETE", "noprompt")
  281. _, err := adapter.raid.term.Run(cmd)
  282. return err
  283. }
  284. type Mpt2SASRaid struct {
  285. term raid.IExecTerm
  286. utility string
  287. adapters []*Mpt2SASRaidAdaptor
  288. }
  289. func NewMpt2SASRaid(term raid.IExecTerm) raid.IRaidDriver {
  290. return &Mpt2SASRaid{
  291. term: term,
  292. adapters: make([]*Mpt2SASRaidAdaptor, 0),
  293. }
  294. }
  295. func (r *Mpt2SASRaid) GetName() string {
  296. return baremetal.DISK_DRIVER_MPT2SAS
  297. }
  298. func (r *Mpt2SASRaid) ParsePhyDevs() error {
  299. if r.modulePCIProbed(raid.MODULE_MPT2SAS) {
  300. r.utility = "/opt/lsi/sas2ircu"
  301. } else if r.modulePCIProbed(raid.MODULE_MPT3SAS) {
  302. r.utility = "/opt/lsi/sas3ircu"
  303. } else {
  304. return fmt.Errorf("Not probe mpt2sas or mpt3sas kernel module")
  305. }
  306. cmd := r.GetCommand("LIST")
  307. ret, err := r.term.Run(cmd)
  308. if err != nil {
  309. return err
  310. }
  311. return r.parseAdapters(ret)
  312. }
  313. func getLineAdapterIndex(line string) int {
  314. dat := regexp.MustCompile(`\s+`).Split(strings.TrimSpace(line), -1)
  315. if len(dat) == 0 {
  316. return -1
  317. }
  318. if regexp.MustCompile(`\d+`).MatchString(dat[0]) {
  319. if !regexp.MustCompile(`^\d+`).MatchString(dat[0]) {
  320. return -1
  321. }
  322. idx, _ := strconv.Atoi(dat[0])
  323. return idx
  324. }
  325. return -1
  326. }
  327. func (r *Mpt2SASRaid) parseAdapters(lines []string) error {
  328. for _, line := range lines {
  329. idx := getLineAdapterIndex(line)
  330. if idx >= 0 {
  331. adapter := newMpt2SASRaidAdaptor(idx, r)
  332. r.adapters = append(r.adapters, adapter)
  333. }
  334. }
  335. for _, adapter := range r.adapters {
  336. if err := adapter.ParsePhyDevs(); err != nil {
  337. return err
  338. }
  339. }
  340. return nil
  341. }
  342. func (r *Mpt2SASRaid) modulePCIProbed(mod string) bool {
  343. cmd := fmt.Sprintf("lspci -k | grep %s", mod)
  344. _, err := r.term.Run(cmd)
  345. return err == nil
  346. }
  347. func (r *Mpt2SASRaid) GetCommand(args ...string) string {
  348. return raid.GetCommand(r.utility, args...)
  349. }
  350. func (r *Mpt2SASRaid) PreBuildRaid(_ []*api.BaremetalDiskConfig, _ int) error {
  351. return nil
  352. }
  353. func (r *Mpt2SASRaid) GetAdapters() []raid.IRaidAdapter {
  354. ret := make([]raid.IRaidAdapter, 0)
  355. for _, a := range r.adapters {
  356. ret = append(ret, a)
  357. }
  358. return ret
  359. }
  360. func (r *Mpt2SASRaid) CleanRaid() error {
  361. return nil
  362. }
  363. func init() {
  364. raid.RegisterDriver(baremetal.DISK_DRIVER_MPT2SAS, NewMpt2SASRaid)
  365. }