for-5.4/libata-2019-09-15
-----BEGIN PGP SIGNATURE----- iQJEBAABCAAuFiEEwPw5LcreJtl1+l5K99NY+ylx4KYFAl1+zyIQHGF4Ym9lQGtl cm5lbC5kawAKCRD301j7KXHgpsNqD/4/JOO/eruWblzrAM+n06HUgyWgherZof7j +Uev4YpBbT74Ll2WcrBB+80XuidQx3fSKZ1Jq7PsBjdSL/oOGAizflLn17gYGgGV cE3gtej1ogDzzdWHgbhnnCvmL6G+8GhXpZeqyOWNhHHDPu6KrieckmUFNUdpjqiC 6VcVrhLngiIa5ZNgOBEa1XKDc/YcbjGEynV0EjufZIELZw8GOIJSONLUW70tUWzv oTEQZNqYAXuyFqcS0qdjRIXHic3tgqMcgkgmxlEtqhlFoF0eIcgkfew2Wz5lyEYT e5jxhv5geY3JZbe0Vxb2FJXAv9O2WCiMc2Lt7PCHcvyfrF1NHHcVrKtP0XLxlkHp HyxiJXqZqNngTKAmm581DvGHojlGqJMliSSBqC3IIAV+Ubeks5S3m3uFcvvfPVAa lKd8eboOrCxqfHsGeaDtA/lDtOYqr5ozXkcBWEPc34SR1O2YDOeL+jUb0Ry/L/dd ZG128z0BpnlLHf6OE+9WCxXJDvVm7PT2i2GS1/JkIhfYw2JzP29vI6dvRF/68x0W aNkNQjbPLUJ/5nilWfPcpfmeCkdZphvkIlyQTVD5wjZbfAoSGqciESG5HExWmDg8 YcgU32YkNX45NzokhuE+ip59V5aT+cHD9mPhHQR5TXIhM9gH0QgXQc7HzRn2WpyB T5V2LWqcxg== =824U -----END PGP SIGNATURE----- Merge tag 'for-5.4/libata-2019-09-15' of git://git.kernel.dk/linux-block Pull libata updates from Jens Axboe: - Kill unused export (Andy) - Use dma_set_mask_and_coherent() throughout (Christoph) - Drop PCS quirk on Denverton, which has different register layout (Dan) - Support non-boot time detection for pata_buddha (Max) * tag 'for-5.4/libata-2019-09-15' of git://git.kernel.dk/linux-block: libata/ahci: Drop PCS quirk for Denverton and beyond ahci: Do not export local variable ahci_em_messages libata: switch remaining drivers to use dma_set_mask_and_coherent sata_sil24: use dma_set_mask_and_coherent sata_qstor: use dma_set_mask_and_coherent sata_nv: use dma_set_mask_and_coherent sata_mv: use dma_set_mask_and_coherent pdc_adma: use dma_set_mask_and_coherent ahci: use dma_set_mask_and_coherent acard_ahci: use dma_set_mask_and_coherent ata/pata_buddha: Probe via modalias instead of initcall
This commit is contained in:
commit
5260c2b863
|
@ -160,37 +160,6 @@ static int acard_ahci_pci_device_resume(struct pci_dev *pdev)
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
static int acard_ahci_configure_dma_masks(struct pci_dev *pdev, int using_dac)
|
|
||||||
{
|
|
||||||
int rc;
|
|
||||||
|
|
||||||
if (using_dac &&
|
|
||||||
!dma_set_mask(&pdev->dev, DMA_BIT_MASK(64))) {
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(64));
|
|
||||||
if (rc) {
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc) {
|
|
||||||
dev_err(&pdev->dev,
|
|
||||||
"64-bit DMA enable failed\n");
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
rc = dma_set_mask(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc) {
|
|
||||||
dev_err(&pdev->dev, "32-bit DMA enable failed\n");
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc) {
|
|
||||||
dev_err(&pdev->dev,
|
|
||||||
"32-bit consistent DMA enable failed\n");
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void acard_ahci_pci_print_info(struct ata_host *host)
|
static void acard_ahci_pci_print_info(struct ata_host *host)
|
||||||
{
|
{
|
||||||
struct pci_dev *pdev = to_pci_dev(host->dev);
|
struct pci_dev *pdev = to_pci_dev(host->dev);
|
||||||
|
@ -471,9 +440,12 @@ static int acard_ahci_init_one(struct pci_dev *pdev, const struct pci_device_id
|
||||||
}
|
}
|
||||||
|
|
||||||
/* initialize adapter */
|
/* initialize adapter */
|
||||||
rc = acard_ahci_configure_dma_masks(pdev, hpriv->cap & HOST_CAP_64);
|
rc = dma_set_mask_and_coherent(&pdev->dev,
|
||||||
if (rc)
|
DMA_BIT_MASK((hpriv->cap & HOST_CAP_64) ? 64 : 32));
|
||||||
|
if (rc) {
|
||||||
|
dev_err(&pdev->dev, "DMA enable failed\n");
|
||||||
return rc;
|
return rc;
|
||||||
|
}
|
||||||
|
|
||||||
rc = ahci_reset_controller(host);
|
rc = ahci_reset_controller(host);
|
||||||
if (rc)
|
if (rc)
|
||||||
|
|
|
@ -65,6 +65,12 @@ enum board_ids {
|
||||||
board_ahci_sb700, /* for SB700 and SB800 */
|
board_ahci_sb700, /* for SB700 and SB800 */
|
||||||
board_ahci_vt8251,
|
board_ahci_vt8251,
|
||||||
|
|
||||||
|
/*
|
||||||
|
* board IDs for Intel chipsets that support more than 6 ports
|
||||||
|
* *and* end up needing the PCS quirk.
|
||||||
|
*/
|
||||||
|
board_ahci_pcs7,
|
||||||
|
|
||||||
/* aliases */
|
/* aliases */
|
||||||
board_ahci_mcp_linux = board_ahci_mcp65,
|
board_ahci_mcp_linux = board_ahci_mcp65,
|
||||||
board_ahci_mcp67 = board_ahci_mcp65,
|
board_ahci_mcp67 = board_ahci_mcp65,
|
||||||
|
@ -220,6 +226,12 @@ static const struct ata_port_info ahci_port_info[] = {
|
||||||
.udma_mask = ATA_UDMA6,
|
.udma_mask = ATA_UDMA6,
|
||||||
.port_ops = &ahci_vt8251_ops,
|
.port_ops = &ahci_vt8251_ops,
|
||||||
},
|
},
|
||||||
|
[board_ahci_pcs7] = {
|
||||||
|
.flags = AHCI_FLAG_COMMON,
|
||||||
|
.pio_mask = ATA_PIO4,
|
||||||
|
.udma_mask = ATA_UDMA6,
|
||||||
|
.port_ops = &ahci_ops,
|
||||||
|
},
|
||||||
};
|
};
|
||||||
|
|
||||||
static const struct pci_device_id ahci_pci_tbl[] = {
|
static const struct pci_device_id ahci_pci_tbl[] = {
|
||||||
|
@ -264,26 +276,26 @@ static const struct pci_device_id ahci_pci_tbl[] = {
|
||||||
{ PCI_VDEVICE(INTEL, 0x3b2b), board_ahci }, /* PCH RAID */
|
{ PCI_VDEVICE(INTEL, 0x3b2b), board_ahci }, /* PCH RAID */
|
||||||
{ PCI_VDEVICE(INTEL, 0x3b2c), board_ahci_mobile }, /* PCH M RAID */
|
{ PCI_VDEVICE(INTEL, 0x3b2c), board_ahci_mobile }, /* PCH M RAID */
|
||||||
{ PCI_VDEVICE(INTEL, 0x3b2f), board_ahci }, /* PCH AHCI */
|
{ PCI_VDEVICE(INTEL, 0x3b2f), board_ahci }, /* PCH AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19b0), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19b0), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19b1), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19b1), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19b2), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19b2), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19b3), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19b3), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19b4), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19b4), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19b5), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19b5), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19b6), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19b6), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19b7), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19b7), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19bE), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19bE), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19bF), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19bF), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19c0), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19c0), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19c1), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19c1), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19c2), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19c2), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19c3), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19c3), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19c4), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19c4), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19c5), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19c5), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19c6), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19c6), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19c7), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19c7), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19cE), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19cE), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x19cF), board_ahci }, /* DNV AHCI */
|
{ PCI_VDEVICE(INTEL, 0x19cF), board_ahci_pcs7 }, /* DNV AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x1c02), board_ahci }, /* CPT AHCI */
|
{ PCI_VDEVICE(INTEL, 0x1c02), board_ahci }, /* CPT AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x1c03), board_ahci_mobile }, /* CPT M AHCI */
|
{ PCI_VDEVICE(INTEL, 0x1c03), board_ahci_mobile }, /* CPT M AHCI */
|
||||||
{ PCI_VDEVICE(INTEL, 0x1c04), board_ahci }, /* CPT RAID */
|
{ PCI_VDEVICE(INTEL, 0x1c04), board_ahci }, /* CPT RAID */
|
||||||
|
@ -623,30 +635,6 @@ static void ahci_pci_save_initial_config(struct pci_dev *pdev,
|
||||||
ahci_save_initial_config(&pdev->dev, hpriv);
|
ahci_save_initial_config(&pdev->dev, hpriv);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int ahci_pci_reset_controller(struct ata_host *host)
|
|
||||||
{
|
|
||||||
struct pci_dev *pdev = to_pci_dev(host->dev);
|
|
||||||
int rc;
|
|
||||||
|
|
||||||
rc = ahci_reset_controller(host);
|
|
||||||
if (rc)
|
|
||||||
return rc;
|
|
||||||
|
|
||||||
if (pdev->vendor == PCI_VENDOR_ID_INTEL) {
|
|
||||||
struct ahci_host_priv *hpriv = host->private_data;
|
|
||||||
u16 tmp16;
|
|
||||||
|
|
||||||
/* configure PCS */
|
|
||||||
pci_read_config_word(pdev, 0x92, &tmp16);
|
|
||||||
if ((tmp16 & hpriv->port_map) != hpriv->port_map) {
|
|
||||||
tmp16 |= hpriv->port_map;
|
|
||||||
pci_write_config_word(pdev, 0x92, tmp16);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void ahci_pci_init_controller(struct ata_host *host)
|
static void ahci_pci_init_controller(struct ata_host *host)
|
||||||
{
|
{
|
||||||
struct ahci_host_priv *hpriv = host->private_data;
|
struct ahci_host_priv *hpriv = host->private_data;
|
||||||
|
@ -849,7 +837,7 @@ static int ahci_pci_device_runtime_resume(struct device *dev)
|
||||||
struct ata_host *host = pci_get_drvdata(pdev);
|
struct ata_host *host = pci_get_drvdata(pdev);
|
||||||
int rc;
|
int rc;
|
||||||
|
|
||||||
rc = ahci_pci_reset_controller(host);
|
rc = ahci_reset_controller(host);
|
||||||
if (rc)
|
if (rc)
|
||||||
return rc;
|
return rc;
|
||||||
ahci_pci_init_controller(host);
|
ahci_pci_init_controller(host);
|
||||||
|
@ -884,7 +872,7 @@ static int ahci_pci_device_resume(struct device *dev)
|
||||||
ahci_mcp89_apple_enable(pdev);
|
ahci_mcp89_apple_enable(pdev);
|
||||||
|
|
||||||
if (pdev->dev.power.power_state.event == PM_EVENT_SUSPEND) {
|
if (pdev->dev.power.power_state.event == PM_EVENT_SUSPEND) {
|
||||||
rc = ahci_pci_reset_controller(host);
|
rc = ahci_reset_controller(host);
|
||||||
if (rc)
|
if (rc)
|
||||||
return rc;
|
return rc;
|
||||||
|
|
||||||
|
@ -901,40 +889,23 @@ static int ahci_pci_device_resume(struct device *dev)
|
||||||
|
|
||||||
static int ahci_configure_dma_masks(struct pci_dev *pdev, int using_dac)
|
static int ahci_configure_dma_masks(struct pci_dev *pdev, int using_dac)
|
||||||
{
|
{
|
||||||
|
const int dma_bits = using_dac ? 64 : 32;
|
||||||
int rc;
|
int rc;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* If the device fixup already set the dma_mask to some non-standard
|
* If the device fixup already set the dma_mask to some non-standard
|
||||||
* value, don't extend it here. This happens on STA2X11, for example.
|
* value, don't extend it here. This happens on STA2X11, for example.
|
||||||
|
*
|
||||||
|
* XXX: manipulating the DMA mask from platform code is completely
|
||||||
|
* bogus, platform code should use dev->bus_dma_mask instead..
|
||||||
*/
|
*/
|
||||||
if (pdev->dma_mask && pdev->dma_mask < DMA_BIT_MASK(32))
|
if (pdev->dma_mask && pdev->dma_mask < DMA_BIT_MASK(32))
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
if (using_dac &&
|
rc = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(dma_bits));
|
||||||
!dma_set_mask(&pdev->dev, DMA_BIT_MASK(64))) {
|
if (rc)
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(64));
|
dev_err(&pdev->dev, "DMA enable failed\n");
|
||||||
if (rc) {
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc) {
|
|
||||||
dev_err(&pdev->dev,
|
|
||||||
"64-bit DMA enable failed\n");
|
|
||||||
return rc;
|
return rc;
|
||||||
}
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
rc = dma_set_mask(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc) {
|
|
||||||
dev_err(&pdev->dev, "32-bit DMA enable failed\n");
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc) {
|
|
||||||
dev_err(&pdev->dev,
|
|
||||||
"32-bit consistent DMA enable failed\n");
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return 0;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void ahci_pci_print_info(struct ata_host *host)
|
static void ahci_pci_print_info(struct ata_host *host)
|
||||||
|
@ -1619,6 +1590,34 @@ static void ahci_update_initial_lpm_policy(struct ata_port *ap,
|
||||||
ap->target_lpm_policy = policy;
|
ap->target_lpm_policy = policy;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void ahci_intel_pcs_quirk(struct pci_dev *pdev, struct ahci_host_priv *hpriv)
|
||||||
|
{
|
||||||
|
const struct pci_device_id *id = pci_match_id(ahci_pci_tbl, pdev);
|
||||||
|
u16 tmp16;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Only apply the 6-port PCS quirk for known legacy platforms.
|
||||||
|
*/
|
||||||
|
if (!id || id->vendor != PCI_VENDOR_ID_INTEL)
|
||||||
|
return;
|
||||||
|
if (((enum board_ids) id->driver_data) < board_ahci_pcs7)
|
||||||
|
return;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* port_map is determined from PORTS_IMPL PCI register which is
|
||||||
|
* implemented as write or write-once register. If the register
|
||||||
|
* isn't programmed, ahci automatically generates it from number
|
||||||
|
* of ports, which is good enough for PCS programming. It is
|
||||||
|
* otherwise expected that platform firmware enables the ports
|
||||||
|
* before the OS boots.
|
||||||
|
*/
|
||||||
|
pci_read_config_word(pdev, PCS_6, &tmp16);
|
||||||
|
if ((tmp16 & hpriv->port_map) != hpriv->port_map) {
|
||||||
|
tmp16 |= hpriv->port_map;
|
||||||
|
pci_write_config_word(pdev, PCS_6, tmp16);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
static int ahci_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
|
static int ahci_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
|
||||||
{
|
{
|
||||||
unsigned int board_id = ent->driver_data;
|
unsigned int board_id = ent->driver_data;
|
||||||
|
@ -1731,6 +1730,12 @@ static int ahci_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
|
||||||
/* save initial config */
|
/* save initial config */
|
||||||
ahci_pci_save_initial_config(pdev, hpriv);
|
ahci_pci_save_initial_config(pdev, hpriv);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* If platform firmware failed to enable ports, try to enable
|
||||||
|
* them here.
|
||||||
|
*/
|
||||||
|
ahci_intel_pcs_quirk(pdev, hpriv);
|
||||||
|
|
||||||
/* prepare host */
|
/* prepare host */
|
||||||
if (hpriv->cap & HOST_CAP_NCQ) {
|
if (hpriv->cap & HOST_CAP_NCQ) {
|
||||||
pi.flags |= ATA_FLAG_NCQ;
|
pi.flags |= ATA_FLAG_NCQ;
|
||||||
|
@ -1840,7 +1845,7 @@ static int ahci_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
|
||||||
if (rc)
|
if (rc)
|
||||||
return rc;
|
return rc;
|
||||||
|
|
||||||
rc = ahci_pci_reset_controller(host);
|
rc = ahci_reset_controller(host);
|
||||||
if (rc)
|
if (rc)
|
||||||
return rc;
|
return rc;
|
||||||
|
|
||||||
|
|
|
@ -247,6 +247,8 @@ enum {
|
||||||
ATA_FLAG_ACPI_SATA | ATA_FLAG_AN,
|
ATA_FLAG_ACPI_SATA | ATA_FLAG_AN,
|
||||||
|
|
||||||
ICH_MAP = 0x90, /* ICH MAP register */
|
ICH_MAP = 0x90, /* ICH MAP register */
|
||||||
|
PCS_6 = 0x92, /* 6 port PCS */
|
||||||
|
PCS_7 = 0x94, /* 7+ port PCS (Denverton) */
|
||||||
|
|
||||||
/* em constants */
|
/* em constants */
|
||||||
EM_MAX_SLOTS = 8,
|
EM_MAX_SLOTS = 8,
|
||||||
|
|
|
@ -175,7 +175,6 @@ struct ata_port_operations ahci_pmp_retry_srst_ops = {
|
||||||
EXPORT_SYMBOL_GPL(ahci_pmp_retry_srst_ops);
|
EXPORT_SYMBOL_GPL(ahci_pmp_retry_srst_ops);
|
||||||
|
|
||||||
static bool ahci_em_messages __read_mostly = true;
|
static bool ahci_em_messages __read_mostly = true;
|
||||||
EXPORT_SYMBOL_GPL(ahci_em_messages);
|
|
||||||
module_param(ahci_em_messages, bool, 0444);
|
module_param(ahci_em_messages, bool, 0444);
|
||||||
/* add other LED protocol types when they become supported */
|
/* add other LED protocol types when they become supported */
|
||||||
MODULE_PARM_DESC(ahci_em_messages,
|
MODULE_PARM_DESC(ahci_em_messages,
|
||||||
|
|
|
@ -3153,15 +3153,9 @@ void ata_pci_bmdma_init(struct ata_host *host)
|
||||||
* ->sff_irq_clear method. Try to initialize bmdma_addr
|
* ->sff_irq_clear method. Try to initialize bmdma_addr
|
||||||
* regardless of dma masks.
|
* regardless of dma masks.
|
||||||
*/
|
*/
|
||||||
rc = dma_set_mask(&pdev->dev, ATA_DMA_MASK);
|
rc = dma_set_mask_and_coherent(&pdev->dev, ATA_DMA_MASK);
|
||||||
if (rc)
|
if (rc)
|
||||||
ata_bmdma_nodma(host, "failed to set dma mask");
|
ata_bmdma_nodma(host, "failed to set dma mask");
|
||||||
if (!rc) {
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, ATA_DMA_MASK);
|
|
||||||
if (rc)
|
|
||||||
ata_bmdma_nodma(host,
|
|
||||||
"failed to set consistent dma mask");
|
|
||||||
}
|
|
||||||
|
|
||||||
/* request and iomap DMA region */
|
/* request and iomap DMA region */
|
||||||
rc = pcim_iomap_regions(pdev, 1 << 4, dev_driver_string(gdev));
|
rc = pcim_iomap_regions(pdev, 1 << 4, dev_driver_string(gdev));
|
||||||
|
|
|
@ -463,12 +463,7 @@ static int atp867x_ata_pci_sff_init_host(struct ata_host *host)
|
||||||
|
|
||||||
atp867x_fixup(host);
|
atp867x_fixup(host);
|
||||||
|
|
||||||
rc = dma_set_mask(&pdev->dev, ATA_DMA_MASK);
|
return dma_set_mask_and_coherent(&pdev->dev, ATA_DMA_MASK);
|
||||||
if (rc)
|
|
||||||
return rc;
|
|
||||||
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, ATA_DMA_MASK);
|
|
||||||
return rc;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static int atp867x_init_one(struct pci_dev *pdev,
|
static int atp867x_init_one(struct pci_dev *pdev,
|
||||||
|
|
|
@ -18,7 +18,9 @@
|
||||||
#include <linux/kernel.h>
|
#include <linux/kernel.h>
|
||||||
#include <linux/libata.h>
|
#include <linux/libata.h>
|
||||||
#include <linux/mm.h>
|
#include <linux/mm.h>
|
||||||
|
#include <linux/mod_devicetable.h>
|
||||||
#include <linux/module.h>
|
#include <linux/module.h>
|
||||||
|
#include <linux/types.h>
|
||||||
#include <linux/zorro.h>
|
#include <linux/zorro.h>
|
||||||
#include <scsi/scsi_cmnd.h>
|
#include <scsi/scsi_cmnd.h>
|
||||||
#include <scsi/scsi_host.h>
|
#include <scsi/scsi_host.h>
|
||||||
|
@ -29,7 +31,7 @@
|
||||||
#include <asm/setup.h>
|
#include <asm/setup.h>
|
||||||
|
|
||||||
#define DRV_NAME "pata_buddha"
|
#define DRV_NAME "pata_buddha"
|
||||||
#define DRV_VERSION "0.1.0"
|
#define DRV_VERSION "0.1.1"
|
||||||
|
|
||||||
#define BUDDHA_BASE1 0x800
|
#define BUDDHA_BASE1 0x800
|
||||||
#define BUDDHA_BASE2 0xa00
|
#define BUDDHA_BASE2 0xa00
|
||||||
|
@ -47,11 +49,11 @@ enum {
|
||||||
BOARD_XSURF
|
BOARD_XSURF
|
||||||
};
|
};
|
||||||
|
|
||||||
static unsigned int buddha_bases[3] __initdata = {
|
static unsigned int buddha_bases[3] = {
|
||||||
BUDDHA_BASE1, BUDDHA_BASE2, BUDDHA_BASE3
|
BUDDHA_BASE1, BUDDHA_BASE2, BUDDHA_BASE3
|
||||||
};
|
};
|
||||||
|
|
||||||
static unsigned int xsurf_bases[2] __initdata = {
|
static unsigned int xsurf_bases[2] = {
|
||||||
XSURF_BASE1, XSURF_BASE2
|
XSURF_BASE1, XSURF_BASE2
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -145,29 +147,20 @@ static struct ata_port_operations pata_xsurf_ops = {
|
||||||
.set_mode = pata_buddha_set_mode,
|
.set_mode = pata_buddha_set_mode,
|
||||||
};
|
};
|
||||||
|
|
||||||
static int __init pata_buddha_init_one(void)
|
static int pata_buddha_probe(struct zorro_dev *z,
|
||||||
|
const struct zorro_device_id *ent)
|
||||||
{
|
{
|
||||||
struct zorro_dev *z = NULL;
|
static const char * const board_name[] = {
|
||||||
|
"Buddha", "Catweasel", "X-Surf"
|
||||||
while ((z = zorro_find_device(ZORRO_WILDCARD, z))) {
|
};
|
||||||
static const char *board_name[]
|
|
||||||
= { "Buddha", "Catweasel", "X-Surf" };
|
|
||||||
struct ata_host *host;
|
struct ata_host *host;
|
||||||
void __iomem *buddha_board;
|
void __iomem *buddha_board;
|
||||||
unsigned long board;
|
unsigned long board;
|
||||||
unsigned int type, nr_ports = 2;
|
unsigned int type = ent->driver_data;
|
||||||
|
unsigned int nr_ports = (type == BOARD_CATWEASEL) ? 3 : 2;
|
||||||
|
void *old_drvdata;
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
if (z->id == ZORRO_PROD_INDIVIDUAL_COMPUTERS_BUDDHA) {
|
|
||||||
type = BOARD_BUDDHA;
|
|
||||||
} else if (z->id == ZORRO_PROD_INDIVIDUAL_COMPUTERS_CATWEASEL) {
|
|
||||||
type = BOARD_CATWEASEL;
|
|
||||||
nr_ports++;
|
|
||||||
} else if (z->id == ZORRO_PROD_INDIVIDUAL_COMPUTERS_X_SURF) {
|
|
||||||
type = BOARD_XSURF;
|
|
||||||
} else
|
|
||||||
continue;
|
|
||||||
|
|
||||||
dev_info(&z->dev, "%s IDE controller\n", board_name[type]);
|
dev_info(&z->dev, "%s IDE controller\n", board_name[type]);
|
||||||
|
|
||||||
board = z->resource.start;
|
board = z->resource.start;
|
||||||
|
@ -176,22 +169,28 @@ static int __init pata_buddha_init_one(void)
|
||||||
if (!devm_request_mem_region(&z->dev,
|
if (!devm_request_mem_region(&z->dev,
|
||||||
board + BUDDHA_BASE1,
|
board + BUDDHA_BASE1,
|
||||||
0x800, DRV_NAME))
|
0x800, DRV_NAME))
|
||||||
continue;
|
return -ENXIO;
|
||||||
} else {
|
} else {
|
||||||
if (!devm_request_mem_region(&z->dev,
|
if (!devm_request_mem_region(&z->dev,
|
||||||
board + XSURF_BASE1,
|
board + XSURF_BASE1,
|
||||||
0x1000, DRV_NAME))
|
0x1000, DRV_NAME))
|
||||||
continue;
|
return -ENXIO;
|
||||||
if (!devm_request_mem_region(&z->dev,
|
if (!devm_request_mem_region(&z->dev,
|
||||||
board + XSURF_BASE2,
|
board + XSURF_BASE2,
|
||||||
0x1000, DRV_NAME))
|
0x1000, DRV_NAME)) {
|
||||||
continue;
|
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Workaround for X-Surf: Save drvdata in case zorro8390 has set it */
|
||||||
|
if (type == BOARD_XSURF)
|
||||||
|
old_drvdata = dev_get_drvdata(&z->dev);
|
||||||
|
|
||||||
/* allocate host */
|
/* allocate host */
|
||||||
host = ata_host_alloc(&z->dev, nr_ports);
|
host = ata_host_alloc(&z->dev, nr_ports);
|
||||||
|
if (type == BOARD_XSURF)
|
||||||
|
dev_set_drvdata(&z->dev, old_drvdata);
|
||||||
if (!host)
|
if (!host)
|
||||||
continue;
|
return -ENXIO;
|
||||||
|
|
||||||
buddha_board = ZTWO_VADDR(board);
|
buddha_board = ZTWO_VADDR(board);
|
||||||
|
|
||||||
|
@ -244,12 +243,55 @@ static int __init pata_buddha_init_one(void)
|
||||||
ata_host_activate(host, IRQ_AMIGA_PORTS, ata_sff_interrupt,
|
ata_host_activate(host, IRQ_AMIGA_PORTS, ata_sff_interrupt,
|
||||||
IRQF_SHARED, &pata_buddha_sht);
|
IRQF_SHARED, &pata_buddha_sht);
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
static void pata_buddha_remove(struct zorro_dev *z)
|
||||||
|
{
|
||||||
|
struct ata_host *host = dev_get_drvdata(&z->dev);
|
||||||
|
|
||||||
|
ata_host_detach(host);
|
||||||
|
}
|
||||||
|
|
||||||
|
static const struct zorro_device_id pata_buddha_zorro_tbl[] = {
|
||||||
|
{ ZORRO_PROD_INDIVIDUAL_COMPUTERS_BUDDHA, BOARD_BUDDHA},
|
||||||
|
{ ZORRO_PROD_INDIVIDUAL_COMPUTERS_CATWEASEL, BOARD_CATWEASEL},
|
||||||
|
{ 0 }
|
||||||
|
};
|
||||||
|
MODULE_DEVICE_TABLE(zorro, pata_buddha_zorro_tbl);
|
||||||
|
|
||||||
|
static struct zorro_driver pata_buddha_driver = {
|
||||||
|
.name = "pata_buddha",
|
||||||
|
.id_table = pata_buddha_zorro_tbl,
|
||||||
|
.probe = pata_buddha_probe,
|
||||||
|
.remove = pata_buddha_remove,
|
||||||
|
};
|
||||||
|
|
||||||
|
/*
|
||||||
|
* We cannot have a modalias for X-Surf boards, as it competes with the
|
||||||
|
* zorro8390 network driver. As a stopgap measure until we have proper
|
||||||
|
* MFD support for this board, we manually attach to it late after Zorro
|
||||||
|
* has enumerated its boards.
|
||||||
|
*/
|
||||||
|
static int __init pata_buddha_late_init(void)
|
||||||
|
{
|
||||||
|
struct zorro_dev *z = NULL;
|
||||||
|
|
||||||
|
/* Auto-bind to regular boards */
|
||||||
|
zorro_register_driver(&pata_buddha_driver);
|
||||||
|
|
||||||
|
/* Manually bind to all X-Surf boards */
|
||||||
|
while ((z = zorro_find_device(ZORRO_PROD_INDIVIDUAL_COMPUTERS_X_SURF, z))) {
|
||||||
|
static struct zorro_device_id xsurf_ent = {
|
||||||
|
ZORRO_PROD_INDIVIDUAL_COMPUTERS_X_SURF, BOARD_XSURF
|
||||||
|
};
|
||||||
|
|
||||||
|
pata_buddha_probe(z, &xsurf_ent);
|
||||||
}
|
}
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
late_initcall(pata_buddha_late_init);
|
||||||
module_init(pata_buddha_init_one);
|
|
||||||
|
|
||||||
MODULE_AUTHOR("Bartlomiej Zolnierkiewicz");
|
MODULE_AUTHOR("Bartlomiej Zolnierkiewicz");
|
||||||
MODULE_DESCRIPTION("low-level driver for Buddha/Catweasel/X-Surf PATA");
|
MODULE_DESCRIPTION("low-level driver for Buddha/Catweasel/X-Surf PATA");
|
||||||
|
|
|
@ -155,14 +155,10 @@ static int cs5520_init_one(struct pci_dev *pdev, const struct pci_device_id *id)
|
||||||
return -ENODEV;
|
return -ENODEV;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (dma_set_mask(&pdev->dev, DMA_BIT_MASK(32))) {
|
if (dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(32))) {
|
||||||
printk(KERN_ERR DRV_NAME ": unable to configure DMA mask.\n");
|
printk(KERN_ERR DRV_NAME ": unable to configure DMA mask.\n");
|
||||||
return -ENODEV;
|
return -ENODEV;
|
||||||
}
|
}
|
||||||
if (dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32))) {
|
|
||||||
printk(KERN_ERR DRV_NAME ": unable to configure consistent DMA mask.\n");
|
|
||||||
return -ENODEV;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Map IO ports and initialize host accordingly */
|
/* Map IO ports and initialize host accordingly */
|
||||||
iomap[0] = devm_ioport_map(&pdev->dev, cmd_port[0], 8);
|
iomap[0] = devm_ioport_map(&pdev->dev, cmd_port[0], 8);
|
||||||
|
|
|
@ -221,10 +221,7 @@ static int hpt3x3_init_one(struct pci_dev *pdev, const struct pci_device_id *id)
|
||||||
if (rc)
|
if (rc)
|
||||||
return rc;
|
return rc;
|
||||||
host->iomap = pcim_iomap_table(pdev);
|
host->iomap = pcim_iomap_table(pdev);
|
||||||
rc = dma_set_mask(&pdev->dev, ATA_DMA_MASK);
|
rc = dma_set_mask_and_coherent(&pdev->dev, ATA_DMA_MASK);
|
||||||
if (rc)
|
|
||||||
return rc;
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, ATA_DMA_MASK);
|
|
||||||
if (rc)
|
if (rc)
|
||||||
return rc;
|
return rc;
|
||||||
|
|
||||||
|
|
|
@ -123,10 +123,7 @@ static int ninja32_init_one(struct pci_dev *dev, const struct pci_device_id *id)
|
||||||
return rc;
|
return rc;
|
||||||
|
|
||||||
host->iomap = pcim_iomap_table(dev);
|
host->iomap = pcim_iomap_table(dev);
|
||||||
rc = dma_set_mask(&dev->dev, ATA_DMA_MASK);
|
rc = dma_set_mask_and_coherent(&dev->dev, ATA_DMA_MASK);
|
||||||
if (rc)
|
|
||||||
return rc;
|
|
||||||
rc = dma_set_coherent_mask(&dev->dev, ATA_DMA_MASK);
|
|
||||||
if (rc)
|
if (rc)
|
||||||
return rc;
|
return rc;
|
||||||
pci_set_master(dev);
|
pci_set_master(dev);
|
||||||
|
|
|
@ -722,11 +722,7 @@ static int pdc2027x_init_one(struct pci_dev *pdev,
|
||||||
return rc;
|
return rc;
|
||||||
host->iomap = pcim_iomap_table(pdev);
|
host->iomap = pcim_iomap_table(pdev);
|
||||||
|
|
||||||
rc = dma_set_mask(&pdev->dev, ATA_DMA_MASK);
|
rc = dma_set_mask_and_coherent(&pdev->dev, ATA_DMA_MASK);
|
||||||
if (rc)
|
|
||||||
return rc;
|
|
||||||
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, ATA_DMA_MASK);
|
|
||||||
if (rc)
|
if (rc)
|
||||||
return rc;
|
return rc;
|
||||||
|
|
||||||
|
|
|
@ -374,10 +374,7 @@ static int sil680_init_one(struct pci_dev *pdev, const struct pci_device_id *id)
|
||||||
host->iomap = pcim_iomap_table(pdev);
|
host->iomap = pcim_iomap_table(pdev);
|
||||||
|
|
||||||
/* Setup DMA masks */
|
/* Setup DMA masks */
|
||||||
rc = dma_set_mask(&pdev->dev, ATA_DMA_MASK);
|
rc = dma_set_mask_and_coherent(&pdev->dev, ATA_DMA_MASK);
|
||||||
if (rc)
|
|
||||||
return rc;
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, ATA_DMA_MASK);
|
|
||||||
if (rc)
|
if (rc)
|
||||||
return rc;
|
return rc;
|
||||||
pci_set_master(pdev);
|
pci_set_master(pdev);
|
||||||
|
|
|
@ -572,23 +572,6 @@ static void adma_host_init(struct ata_host *host, unsigned int chip_id)
|
||||||
adma_reset_engine(host->ports[port_no]);
|
adma_reset_engine(host->ports[port_no]);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int adma_set_dma_masks(struct pci_dev *pdev, void __iomem *mmio_base)
|
|
||||||
{
|
|
||||||
int rc;
|
|
||||||
|
|
||||||
rc = dma_set_mask(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc) {
|
|
||||||
dev_err(&pdev->dev, "32-bit DMA enable failed\n");
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc) {
|
|
||||||
dev_err(&pdev->dev, "32-bit consistent DMA enable failed\n");
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
static int adma_ata_init_one(struct pci_dev *pdev,
|
static int adma_ata_init_one(struct pci_dev *pdev,
|
||||||
const struct pci_device_id *ent)
|
const struct pci_device_id *ent)
|
||||||
{
|
{
|
||||||
|
@ -619,9 +602,11 @@ static int adma_ata_init_one(struct pci_dev *pdev,
|
||||||
host->iomap = pcim_iomap_table(pdev);
|
host->iomap = pcim_iomap_table(pdev);
|
||||||
mmio_base = host->iomap[ADMA_MMIO_BAR];
|
mmio_base = host->iomap[ADMA_MMIO_BAR];
|
||||||
|
|
||||||
rc = adma_set_dma_masks(pdev, mmio_base);
|
rc = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(32));
|
||||||
if (rc)
|
if (rc) {
|
||||||
|
dev_err(&pdev->dev, "32-bit DMA enable failed\n");
|
||||||
return rc;
|
return rc;
|
||||||
|
}
|
||||||
|
|
||||||
for (port_no = 0; port_no < ADMA_PORTS; ++port_no) {
|
for (port_no = 0; port_no < ADMA_PORTS; ++port_no) {
|
||||||
struct ata_port *ap = host->ports[port_no];
|
struct ata_port *ap = host->ports[port_no];
|
||||||
|
|
|
@ -862,18 +862,12 @@ static int inic_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Set dma_mask. This devices doesn't support 64bit addressing. */
|
/* Set dma_mask. This devices doesn't support 64bit addressing. */
|
||||||
rc = dma_set_mask(&pdev->dev, DMA_BIT_MASK(32));
|
rc = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(32));
|
||||||
if (rc) {
|
if (rc) {
|
||||||
dev_err(&pdev->dev, "32-bit DMA enable failed\n");
|
dev_err(&pdev->dev, "32-bit DMA enable failed\n");
|
||||||
return rc;
|
return rc;
|
||||||
}
|
}
|
||||||
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc) {
|
|
||||||
dev_err(&pdev->dev, "32-bit consistent DMA enable failed\n");
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
|
|
||||||
rc = init_controller(hpriv->mmio_base, hpriv->cached_hctl);
|
rc = init_controller(hpriv->mmio_base, hpriv->cached_hctl);
|
||||||
if (rc) {
|
if (rc) {
|
||||||
dev_err(&pdev->dev, "failed to initialize controller\n");
|
dev_err(&pdev->dev, "failed to initialize controller\n");
|
||||||
|
|
|
@ -4314,38 +4314,6 @@ static struct pci_driver mv_pci_driver = {
|
||||||
|
|
||||||
};
|
};
|
||||||
|
|
||||||
/* move to PCI layer or libata core? */
|
|
||||||
static int pci_go_64(struct pci_dev *pdev)
|
|
||||||
{
|
|
||||||
int rc;
|
|
||||||
|
|
||||||
if (!dma_set_mask(&pdev->dev, DMA_BIT_MASK(64))) {
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(64));
|
|
||||||
if (rc) {
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc) {
|
|
||||||
dev_err(&pdev->dev,
|
|
||||||
"64-bit DMA enable failed\n");
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
rc = dma_set_mask(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc) {
|
|
||||||
dev_err(&pdev->dev, "32-bit DMA enable failed\n");
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc) {
|
|
||||||
dev_err(&pdev->dev,
|
|
||||||
"32-bit consistent DMA enable failed\n");
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* mv_print_info - Dump key info to kernel log for perusal.
|
* mv_print_info - Dump key info to kernel log for perusal.
|
||||||
* @host: ATA host to print info about
|
* @host: ATA host to print info about
|
||||||
|
@ -4430,9 +4398,11 @@ static int mv_pci_init_one(struct pci_dev *pdev,
|
||||||
host->iomap = pcim_iomap_table(pdev);
|
host->iomap = pcim_iomap_table(pdev);
|
||||||
hpriv->base = host->iomap[MV_PRIMARY_BAR];
|
hpriv->base = host->iomap[MV_PRIMARY_BAR];
|
||||||
|
|
||||||
rc = pci_go_64(pdev);
|
rc = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64));
|
||||||
if (rc)
|
if (rc) {
|
||||||
|
dev_err(&pdev->dev, "DMA enable failed\n");
|
||||||
return rc;
|
return rc;
|
||||||
|
}
|
||||||
|
|
||||||
rc = mv_create_dma_pools(hpriv, &pdev->dev);
|
rc = mv_create_dma_pools(hpriv, &pdev->dev);
|
||||||
if (rc)
|
if (rc)
|
||||||
|
|
|
@ -1122,14 +1122,10 @@ static int nv_adma_port_start(struct ata_port *ap)
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Now that the legacy PRD and padding buffer are allocated we can
|
* Now that the legacy PRD and padding buffer are allocated we can
|
||||||
* try to raise the DMA mask to allocate the CPB/APRD table.
|
* raise the DMA mask to allocate the CPB/APRD table.
|
||||||
*/
|
*/
|
||||||
rc = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64));
|
dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64));
|
||||||
if (rc) {
|
|
||||||
rc = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc)
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
pp->adma_dma_mask = *dev->dma_mask;
|
pp->adma_dma_mask = *dev->dma_mask;
|
||||||
|
|
||||||
mem = dmam_alloc_coherent(dev, NV_ADMA_PORT_PRIV_DMA_SZ,
|
mem = dmam_alloc_coherent(dev, NV_ADMA_PORT_PRIV_DMA_SZ,
|
||||||
|
|
|
@ -1230,10 +1230,7 @@ static int pdc_ata_init_one(struct pci_dev *pdev,
|
||||||
/* initialize adapter */
|
/* initialize adapter */
|
||||||
pdc_host_init(host);
|
pdc_host_init(host);
|
||||||
|
|
||||||
rc = dma_set_mask(&pdev->dev, ATA_DMA_MASK);
|
rc = dma_set_mask_and_coherent(&pdev->dev, ATA_DMA_MASK);
|
||||||
if (rc)
|
|
||||||
return rc;
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, ATA_DMA_MASK);
|
|
||||||
if (rc)
|
if (rc)
|
||||||
return rc;
|
return rc;
|
||||||
|
|
||||||
|
|
|
@ -537,33 +537,13 @@ static void qs_host_init(struct ata_host *host, unsigned int chip_id)
|
||||||
static int qs_set_dma_masks(struct pci_dev *pdev, void __iomem *mmio_base)
|
static int qs_set_dma_masks(struct pci_dev *pdev, void __iomem *mmio_base)
|
||||||
{
|
{
|
||||||
u32 bus_info = readl(mmio_base + QS_HID_HPHY);
|
u32 bus_info = readl(mmio_base + QS_HID_HPHY);
|
||||||
int rc, have_64bit_bus = (bus_info & QS_HPHY_64BIT);
|
int dma_bits = (bus_info & QS_HPHY_64BIT) ? 64 : 32;
|
||||||
|
int rc;
|
||||||
|
|
||||||
if (have_64bit_bus &&
|
rc = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(dma_bits));
|
||||||
!dma_set_mask(&pdev->dev, DMA_BIT_MASK(64))) {
|
if (rc)
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(64));
|
dev_err(&pdev->dev, "%d-bit DMA enable failed\n", dma_bits);
|
||||||
if (rc) {
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc) {
|
|
||||||
dev_err(&pdev->dev,
|
|
||||||
"64-bit DMA enable failed\n");
|
|
||||||
return rc;
|
return rc;
|
||||||
}
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
rc = dma_set_mask(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc) {
|
|
||||||
dev_err(&pdev->dev, "32-bit DMA enable failed\n");
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc) {
|
|
||||||
dev_err(&pdev->dev,
|
|
||||||
"32-bit consistent DMA enable failed\n");
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return 0;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static int qs_ata_init_one(struct pci_dev *pdev,
|
static int qs_ata_init_one(struct pci_dev *pdev,
|
||||||
|
|
|
@ -757,10 +757,7 @@ static int sil_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
|
||||||
return rc;
|
return rc;
|
||||||
host->iomap = pcim_iomap_table(pdev);
|
host->iomap = pcim_iomap_table(pdev);
|
||||||
|
|
||||||
rc = dma_set_mask(&pdev->dev, ATA_DMA_MASK);
|
rc = dma_set_mask_and_coherent(&pdev->dev, ATA_DMA_MASK);
|
||||||
if (rc)
|
|
||||||
return rc;
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, ATA_DMA_MASK);
|
|
||||||
if (rc)
|
if (rc)
|
||||||
return rc;
|
return rc;
|
||||||
|
|
||||||
|
|
|
@ -1301,29 +1301,11 @@ static int sil24_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
|
||||||
host->iomap = iomap;
|
host->iomap = iomap;
|
||||||
|
|
||||||
/* configure and activate the device */
|
/* configure and activate the device */
|
||||||
if (!dma_set_mask(&pdev->dev, DMA_BIT_MASK(64))) {
|
rc = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64));
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(64));
|
|
||||||
if (rc) {
|
if (rc) {
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
|
dev_err(&pdev->dev, "DMA enable failed\n");
|
||||||
if (rc) {
|
|
||||||
dev_err(&pdev->dev,
|
|
||||||
"64-bit DMA enable failed\n");
|
|
||||||
return rc;
|
return rc;
|
||||||
}
|
}
|
||||||
}
|
|
||||||
} else {
|
|
||||||
rc = dma_set_mask(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc) {
|
|
||||||
dev_err(&pdev->dev, "32-bit DMA enable failed\n");
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc) {
|
|
||||||
dev_err(&pdev->dev,
|
|
||||||
"32-bit consistent DMA enable failed\n");
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Set max read request size to 4096. This slightly increases
|
/* Set max read request size to 4096. This slightly increases
|
||||||
* write throughput for pci-e variants.
|
* write throughput for pci-e variants.
|
||||||
|
|
|
@ -471,10 +471,7 @@ static int k2_sata_init_one(struct pci_dev *pdev, const struct pci_device_id *en
|
||||||
ata_port_pbar_desc(ap, 5, offset, "port");
|
ata_port_pbar_desc(ap, 5, offset, "port");
|
||||||
}
|
}
|
||||||
|
|
||||||
rc = dma_set_mask(&pdev->dev, ATA_DMA_MASK);
|
rc = dma_set_mask_and_coherent(&pdev->dev, ATA_DMA_MASK);
|
||||||
if (rc)
|
|
||||||
return rc;
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, ATA_DMA_MASK);
|
|
||||||
if (rc)
|
if (rc)
|
||||||
return rc;
|
return rc;
|
||||||
|
|
||||||
|
|
|
@ -1470,10 +1470,7 @@ static int pdc_sata_init_one(struct pci_dev *pdev,
|
||||||
}
|
}
|
||||||
|
|
||||||
/* configure and activate */
|
/* configure and activate */
|
||||||
rc = dma_set_mask(&pdev->dev, ATA_DMA_MASK);
|
rc = dma_set_mask_and_coherent(&pdev->dev, ATA_DMA_MASK);
|
||||||
if (rc)
|
|
||||||
return rc;
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, ATA_DMA_MASK);
|
|
||||||
if (rc)
|
if (rc)
|
||||||
return rc;
|
return rc;
|
||||||
|
|
||||||
|
|
|
@ -505,14 +505,7 @@ static int vt6421_prepare_host(struct pci_dev *pdev, struct ata_host **r_host)
|
||||||
for (i = 0; i < host->n_ports; i++)
|
for (i = 0; i < host->n_ports; i++)
|
||||||
vt6421_init_addrs(host->ports[i]);
|
vt6421_init_addrs(host->ports[i]);
|
||||||
|
|
||||||
rc = dma_set_mask(&pdev->dev, ATA_DMA_MASK);
|
return dma_set_mask_and_coherent(&pdev->dev, ATA_DMA_MASK);
|
||||||
if (rc)
|
|
||||||
return rc;
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, ATA_DMA_MASK);
|
|
||||||
if (rc)
|
|
||||||
return rc;
|
|
||||||
|
|
||||||
return 0;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static int vt8251_prepare_host(struct pci_dev *pdev, struct ata_host **r_host)
|
static int vt8251_prepare_host(struct pci_dev *pdev, struct ata_host **r_host)
|
||||||
|
|
|
@ -371,10 +371,7 @@ static int vsc_sata_init_one(struct pci_dev *pdev,
|
||||||
/*
|
/*
|
||||||
* Use 32 bit DMA mask, because 64 bit address support is poor.
|
* Use 32 bit DMA mask, because 64 bit address support is poor.
|
||||||
*/
|
*/
|
||||||
rc = dma_set_mask(&pdev->dev, DMA_BIT_MASK(32));
|
rc = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(32));
|
||||||
if (rc)
|
|
||||||
return rc;
|
|
||||||
rc = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
|
|
||||||
if (rc)
|
if (rc)
|
||||||
return rc;
|
return rc;
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue