123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488148914901491149214931494149514961497149814991500150115021503150415051506150715081509151015111512151315141515151615171518151915201521152215231524152515261527152815291530153115321533153415351536153715381539154015411542154315441545154615471548154915501551155215531554155515561557155815591560156115621563156415651566156715681569157015711572157315741575157615771578157915801581158215831584158515861587158815891590159115921593159415951596159715981599160016011602160316041605160616071608160916101611161216131614161516161617161816191620162116221623162416251626162716281629163016311632163316341635163616371638163916401641164216431644164516461647164816491650165116521653165416551656165716581659166016611662166316641665166616671668166916701671167216731674167516761677167816791680168116821683168416851686168716881689169016911692169316941695169616971698169917001701170217031704170517061707170817091710171117121713171417151716171717181719172017211722172317241725172617271728172917301731173217331734173517361737173817391740174117421743174417451746174717481749175017511752175317541755175617571758175917601761176217631764176517661767176817691770177117721773177417751776177717781779178017811782178317841785178617871788178917901791179217931794179517961797179817991800180118021803180418051806180718081809181018111812181318141815181618171818181918201821182218231824182518261827182818291830183118321833183418351836183718381839184018411842184318441845184618471848184918501851185218531854185518561857185818591860186118621863186418651866186718681869187018711872187318741875187618771878187918801881188218831884188518861887188818891890189118921893189418951896189718981899190019011902190319041905190619071908190919101911191219131914191519161917191819191920192119221923192419251926192719281929193019311932193319341935193619371938193919401941194219431944194519461947194819491950195119521953195419551956195719581959196019611962196319641965196619671968196919701971197219731974197519761977197819791980198119821983198419851986198719881989199019911992199319941995199619971998199920002001200220032004200520062007200820092010201120122013201420152016201720182019202020212022202320242025202620272028202920302031203220332034203520362037203820392040204120422043204420452046204720482049205020512052205320542055205620572058205920602061206220632064206520662067206820692070207120722073207420752076207720782079208020812082208320842085208620872088208920902091209220932094209520962097209820992100210121022103210421052106210721082109211021112112211321142115211621172118211921202121212221232124212521262127212821292130213121322133213421352136213721382139214021412142214321442145214621472148214921502151215221532154215521562157215821592160216121622163216421652166216721682169217021712172217321742175217621772178217921802181218221832184218521862187218821892190219121922193219421952196219721982199220022012202220322042205220622072208220922102211221222132214221522162217221822192220222122222223222422252226222722282229223022312232223322342235223622372238223922402241224222432244224522462247224822492250225122522253225422552256225722582259226022612262226322642265226622672268226922702271227222732274227522762277227822792280228122822283228422852286228722882289229022912292229322942295229622972298229923002301230223032304230523062307230823092310231123122313231423152316231723182319232023212322232323242325232623272328232923302331233223332334233523362337233823392340234123422343234423452346234723482349235023512352235323542355235623572358235923602361236223632364236523662367236823692370237123722373237423752376237723782379238023812382238323842385238623872388238923902391239223932394239523962397239823992400240124022403240424052406240724082409241024112412241324142415241624172418241924202421242224232424242524262427242824292430243124322433243424352436243724382439244024412442244324442445244624472448244924502451245224532454245524562457245824592460246124622463246424652466246724682469247024712472247324742475247624772478247924802481248224832484248524862487248824892490249124922493249424952496249724982499250025012502250325042505250625072508250925102511251225132514251525162517251825192520252125222523252425252526252725282529253025312532253325342535253625372538253925402541254225432544254525462547254825492550255125522553255425552556255725582559256025612562256325642565256625672568256925702571257225732574257525762577257825792580258125822583258425852586258725882589259025912592259325942595259625972598259926002601260226032604260526062607260826092610261126122613261426152616261726182619262026212622262326242625262626272628262926302631263226332634263526362637263826392640264126422643264426452646264726482649265026512652265326542655265626572658265926602661266226632664266526662667266826692670267126722673267426752676267726782679268026812682268326842685268626872688268926902691269226932694269526962697269826992700270127022703270427052706270727082709271027112712271327142715 |
- From 2abe91b53ca4d2528ef1fc9c44c6e69f8c805776 Mon Sep 17 00:00:00 2001
- From: John Crispin <blogic@openwrt.org>
- Date: Wed, 18 Nov 2015 03:12:19 +0100
- Subject: [PATCH 501/513] net-next: mediatek: add the drivers core files
- This patch adds the main chunk of the driver. The ethernet core is used in all
- of the Mediatek/Ralink Wireless SoCs. Over the years we have seen verious
- changes to
- * the register layout
- * the type of ports (single/dual gbit, internal FE/Gbit switch)
- * dma engine
- and new offloading features were added, such as
- * checksum
- * vlan tx/rx
- * gso
- * lro
- However the core functionality has remained the sama allowing us to use the
- same core for all SoCs.
- The abstraction for the various SoCs uses the typical ops struct pattern which
- allows us to extend or override the cores functionality depending on which SoC
- we are on. The code to bring up the switches and external ports has also been
- split into separate files.
- Signed-off-by: John Crispin <blogic@openwrt.org>
- Signed-off-by: Felix Fietkau <nbd@nbd.name>
- Signed-off-by: Michael Lee <igvtee@gmail.com>
- ---
- drivers/net/ethernet/mediatek/ethtool.c | 235 ++++
- drivers/net/ethernet/mediatek/ethtool.h | 22 +
- drivers/net/ethernet/mediatek/mdio.c | 258 +++++
- drivers/net/ethernet/mediatek/mdio.h | 27 +
- drivers/net/ethernet/mediatek/mtk_eth_soc.c | 1607 +++++++++++++++++++++++++++
- drivers/net/ethernet/mediatek/mtk_eth_soc.h | 522 +++++++++
- 6 files changed, 2671 insertions(+)
- create mode 100644 drivers/net/ethernet/mediatek/ethtool.c
- create mode 100644 drivers/net/ethernet/mediatek/ethtool.h
- create mode 100644 drivers/net/ethernet/mediatek/mdio.c
- create mode 100644 drivers/net/ethernet/mediatek/mdio.h
- create mode 100644 drivers/net/ethernet/mediatek/mtk_eth_soc.c
- create mode 100644 drivers/net/ethernet/mediatek/mtk_eth_soc.h
- --- /dev/null
- +++ b/drivers/net/ethernet/mediatek/ethtool.c
- @@ -0,0 +1,235 @@
- +/* This program is free software; you can redistribute it and/or modify
- + * it under the terms of the GNU General Public License as published by
- + * the Free Software Foundation; version 2 of the License
- + *
- + * This program is distributed in the hope that it will be useful,
- + * but WITHOUT ANY WARRANTY; without even the implied warranty of
- + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- + * GNU General Public License for more details.
- + *
- + * Copyright (C) 2009-2015 John Crispin <blogic@openwrt.org>
- + * Copyright (C) 2009-2015 Felix Fietkau <nbd@nbd.name>
- + * Copyright (C) 2013-2015 Michael Lee <igvtee@gmail.com>
- + */
- +
- +#include "mtk_eth_soc.h"
- +
- +static const char fe_gdma_str[][ETH_GSTRING_LEN] = {
- +#define _FE(x...) # x,
- +FE_STAT_REG_DECLARE
- +#undef _FE
- +};
- +
- +static int fe_get_settings(struct net_device *dev,
- + struct ethtool_cmd *cmd)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- + int err;
- +
- + if (!priv->phy_dev)
- + goto out_gset;
- +
- + if (priv->phy_flags == FE_PHY_FLAG_ATTACH) {
- + err = phy_read_status(priv->phy_dev);
- + if (err)
- + goto out_gset;
- + }
- +
- + return phy_ethtool_gset(priv->phy_dev, cmd);
- +
- +out_gset:
- + return -ENODEV;
- +}
- +
- +static int fe_set_settings(struct net_device *dev,
- + struct ethtool_cmd *cmd)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- +
- + if (!priv->phy_dev)
- + goto out_sset;
- +
- + if (cmd->phy_address != priv->phy_dev->addr) {
- + if (priv->phy->phy_node[cmd->phy_address]) {
- + priv->phy_dev = priv->phy->phy[cmd->phy_address];
- + priv->phy_flags = FE_PHY_FLAG_PORT;
- + } else if (priv->mii_bus &&
- + priv->mii_bus->phy_map[cmd->phy_address]) {
- + priv->phy_dev =
- + priv->mii_bus->phy_map[cmd->phy_address];
- + priv->phy_flags = FE_PHY_FLAG_ATTACH;
- + } else {
- + goto out_sset;
- + }
- + }
- +
- + return phy_ethtool_sset(priv->phy_dev, cmd);
- +
- +out_sset:
- + return -ENODEV;
- +}
- +
- +static void fe_get_drvinfo(struct net_device *dev,
- + struct ethtool_drvinfo *info)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- + struct fe_soc_data *soc = priv->soc;
- +
- + strlcpy(info->driver, priv->device->driver->name, sizeof(info->driver));
- + strlcpy(info->version, MTK_FE_DRV_VERSION, sizeof(info->version));
- + strlcpy(info->bus_info, dev_name(priv->device), sizeof(info->bus_info));
- +
- + if (soc->reg_table[FE_REG_FE_COUNTER_BASE])
- + info->n_stats = ARRAY_SIZE(fe_gdma_str);
- +}
- +
- +static u32 fe_get_msglevel(struct net_device *dev)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- +
- + return priv->msg_enable;
- +}
- +
- +static void fe_set_msglevel(struct net_device *dev, u32 value)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- +
- + priv->msg_enable = value;
- +}
- +
- +static int fe_nway_reset(struct net_device *dev)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- +
- + if (!priv->phy_dev)
- + goto out_nway_reset;
- +
- + return genphy_restart_aneg(priv->phy_dev);
- +
- +out_nway_reset:
- + return -EOPNOTSUPP;
- +}
- +
- +static u32 fe_get_link(struct net_device *dev)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- + int err;
- +
- + if (!priv->phy_dev)
- + goto out_get_link;
- +
- + if (priv->phy_flags == FE_PHY_FLAG_ATTACH) {
- + err = genphy_update_link(priv->phy_dev);
- + if (err)
- + goto out_get_link;
- + }
- +
- + return priv->phy_dev->link;
- +
- +out_get_link:
- + return ethtool_op_get_link(dev);
- +}
- +
- +static int fe_set_ringparam(struct net_device *dev,
- + struct ethtool_ringparam *ring)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- +
- + if ((ring->tx_pending < 2) ||
- + (ring->rx_pending < 2) ||
- + (ring->rx_pending > MAX_DMA_DESC) ||
- + (ring->tx_pending > MAX_DMA_DESC))
- + return -EINVAL;
- +
- + dev->netdev_ops->ndo_stop(dev);
- +
- + priv->tx_ring.tx_ring_size = BIT(fls(ring->tx_pending) - 1);
- + priv->rx_ring.rx_ring_size = BIT(fls(ring->rx_pending) - 1);
- +
- + dev->netdev_ops->ndo_open(dev);
- +
- + return 0;
- +}
- +
- +static void fe_get_ringparam(struct net_device *dev,
- + struct ethtool_ringparam *ring)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- +
- + ring->rx_max_pending = MAX_DMA_DESC;
- + ring->tx_max_pending = MAX_DMA_DESC;
- + ring->rx_pending = priv->rx_ring.rx_ring_size;
- + ring->tx_pending = priv->tx_ring.tx_ring_size;
- +}
- +
- +static void fe_get_strings(struct net_device *dev, u32 stringset, u8 *data)
- +{
- + switch (stringset) {
- + case ETH_SS_STATS:
- + memcpy(data, *fe_gdma_str, sizeof(fe_gdma_str));
- + break;
- + }
- +}
- +
- +static int fe_get_sset_count(struct net_device *dev, int sset)
- +{
- + switch (sset) {
- + case ETH_SS_STATS:
- + return ARRAY_SIZE(fe_gdma_str);
- + default:
- + return -EOPNOTSUPP;
- + }
- +}
- +
- +static void fe_get_ethtool_stats(struct net_device *dev,
- + struct ethtool_stats *stats, u64 *data)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- + struct fe_hw_stats *hwstats = priv->hw_stats;
- + u64 *data_src, *data_dst;
- + unsigned int start;
- + int i;
- +
- + if (netif_running(dev) && netif_device_present(dev)) {
- + if (spin_trylock(&hwstats->stats_lock)) {
- + fe_stats_update(priv);
- + spin_unlock(&hwstats->stats_lock);
- + }
- + }
- +
- + do {
- + data_src = &hwstats->tx_bytes;
- + data_dst = data;
- + start = u64_stats_fetch_begin_irq(&hwstats->syncp);
- +
- + for (i = 0; i < ARRAY_SIZE(fe_gdma_str); i++)
- + *data_dst++ = *data_src++;
- +
- + } while (u64_stats_fetch_retry_irq(&hwstats->syncp, start));
- +}
- +
- +static struct ethtool_ops fe_ethtool_ops = {
- + .get_settings = fe_get_settings,
- + .set_settings = fe_set_settings,
- + .get_drvinfo = fe_get_drvinfo,
- + .get_msglevel = fe_get_msglevel,
- + .set_msglevel = fe_set_msglevel,
- + .nway_reset = fe_nway_reset,
- + .get_link = fe_get_link,
- + .set_ringparam = fe_set_ringparam,
- + .get_ringparam = fe_get_ringparam,
- +};
- +
- +void fe_set_ethtool_ops(struct net_device *netdev)
- +{
- + struct fe_priv *priv = netdev_priv(netdev);
- + struct fe_soc_data *soc = priv->soc;
- +
- + if (soc->reg_table[FE_REG_FE_COUNTER_BASE]) {
- + fe_ethtool_ops.get_strings = fe_get_strings;
- + fe_ethtool_ops.get_sset_count = fe_get_sset_count;
- + fe_ethtool_ops.get_ethtool_stats = fe_get_ethtool_stats;
- + }
- +
- + netdev->ethtool_ops = &fe_ethtool_ops;
- +}
- --- /dev/null
- +++ b/drivers/net/ethernet/mediatek/ethtool.h
- @@ -0,0 +1,22 @@
- +/* This program is free software; you can redistribute it and/or modify
- + * it under the terms of the GNU General Public License as published by
- + * the Free Software Foundation; version 2 of the License
- + *
- + * This program is distributed in the hope that it will be useful,
- + * but WITHOUT ANY WARRANTY; without even the implied warranty of
- + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- + * GNU General Public License for more details.
- + *
- + * Copyright (C) 2009-2015 John Crispin <blogic@openwrt.org>
- + * Copyright (C) 2009-2015 Felix Fietkau <nbd@nbd.name>
- + * Copyright (C) 2013-2015 Michael Lee <igvtee@gmail.com>
- + */
- +
- +#ifndef FE_ETHTOOL_H
- +#define FE_ETHTOOL_H
- +
- +#include <linux/ethtool.h>
- +
- +void fe_set_ethtool_ops(struct net_device *netdev);
- +
- +#endif /* FE_ETHTOOL_H */
- --- /dev/null
- +++ b/drivers/net/ethernet/mediatek/mdio.c
- @@ -0,0 +1,258 @@
- +/* This program is free software; you can redistribute it and/or modify
- + * it under the terms of the GNU General Public License as published by
- + * the Free Software Foundation; version 2 of the License
- + *
- + * Copyright (C) 2009-2015 John Crispin <blogic@openwrt.org>
- + * Copyright (C) 2009-2015 Felix Fietkau <nbd@nbd.name>
- + * Copyright (C) 2013-2015 Michael Lee <igvtee@gmail.com>
- + */
- +
- +#include <linux/module.h>
- +#include <linux/kernel.h>
- +#include <linux/phy.h>
- +#include <linux/of_net.h>
- +#include <linux/of_mdio.h>
- +
- +#include "mtk_eth_soc.h"
- +#include "mdio.h"
- +
- +static int fe_mdio_reset(struct mii_bus *bus)
- +{
- + /* TODO */
- + return 0;
- +}
- +
- +static void fe_phy_link_adjust(struct net_device *dev)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- + unsigned long flags;
- + int i;
- +
- + spin_lock_irqsave(&priv->phy->lock, flags);
- + for (i = 0; i < 8; i++) {
- + if (priv->phy->phy_node[i]) {
- + struct phy_device *phydev = priv->phy->phy[i];
- + int status_change = 0;
- +
- + if (phydev->link)
- + if (priv->phy->duplex[i] != phydev->duplex ||
- + priv->phy->speed[i] != phydev->speed)
- + status_change = 1;
- +
- + if (phydev->link != priv->link[i])
- + status_change = 1;
- +
- + switch (phydev->speed) {
- + case SPEED_1000:
- + case SPEED_100:
- + case SPEED_10:
- + priv->link[i] = phydev->link;
- + priv->phy->duplex[i] = phydev->duplex;
- + priv->phy->speed[i] = phydev->speed;
- +
- + if (status_change &&
- + priv->soc->mdio_adjust_link)
- + priv->soc->mdio_adjust_link(priv, i);
- + break;
- + }
- + }
- + }
- +}
- +
- +int fe_connect_phy_node(struct fe_priv *priv, struct device_node *phy_node)
- +{
- + const __be32 *_port = NULL;
- + struct phy_device *phydev;
- + int phy_mode, port;
- +
- + _port = of_get_property(phy_node, "reg", NULL);
- +
- + if (!_port || (be32_to_cpu(*_port) >= 0x20)) {
- + pr_err("%s: invalid port id\n", phy_node->name);
- + return -EINVAL;
- + }
- + port = be32_to_cpu(*_port);
- + phy_mode = of_get_phy_mode(phy_node);
- + if (phy_mode < 0) {
- + dev_err(priv->device, "incorrect phy-mode %d\n", phy_mode);
- + priv->phy->phy_node[port] = NULL;
- + return -EINVAL;
- + }
- +
- + phydev = of_phy_connect(priv->netdev, phy_node, fe_phy_link_adjust,
- + 0, phy_mode);
- + if (IS_ERR(phydev)) {
- + dev_err(priv->device, "could not connect to PHY\n");
- + priv->phy->phy_node[port] = NULL;
- + return PTR_ERR(phydev);
- + }
- +
- + phydev->supported &= PHY_GBIT_FEATURES;
- + phydev->advertising = phydev->supported;
- +
- + dev_info(priv->device,
- + "connected port %d to PHY at %s [uid=%08x, driver=%s]\n",
- + port, dev_name(&phydev->dev), phydev->phy_id,
- + phydev->drv->name);
- +
- + priv->phy->phy[port] = phydev;
- + priv->link[port] = 0;
- +
- + return 0;
- +}
- +
- +static void phy_init(struct fe_priv *priv, struct phy_device *phy)
- +{
- + phy_attach(priv->netdev, dev_name(&phy->dev), PHY_INTERFACE_MODE_MII);
- +
- + phy->autoneg = AUTONEG_ENABLE;
- + phy->speed = 0;
- + phy->duplex = 0;
- + phy->supported &= PHY_BASIC_FEATURES;
- + phy->advertising = phy->supported | ADVERTISED_Autoneg;
- +
- + phy_start_aneg(phy);
- +}
- +
- +static int fe_phy_connect(struct fe_priv *priv)
- +{
- + int i;
- +
- + for (i = 0; i < 8; i++) {
- + if (priv->phy->phy_node[i]) {
- + if (!priv->phy_dev) {
- + priv->phy_dev = priv->phy->phy[i];
- + priv->phy_flags = FE_PHY_FLAG_PORT;
- + }
- + } else if (priv->mii_bus && priv->mii_bus->phy_map[i]) {
- + phy_init(priv, priv->mii_bus->phy_map[i]);
- + if (!priv->phy_dev) {
- + priv->phy_dev = priv->mii_bus->phy_map[i];
- + priv->phy_flags = FE_PHY_FLAG_ATTACH;
- + }
- + }
- + }
- +
- + return 0;
- +}
- +
- +static void fe_phy_disconnect(struct fe_priv *priv)
- +{
- + unsigned long flags;
- + int i;
- +
- + for (i = 0; i < 8; i++)
- + if (priv->phy->phy_fixed[i]) {
- + spin_lock_irqsave(&priv->phy->lock, flags);
- + priv->link[i] = 0;
- + if (priv->soc->mdio_adjust_link)
- + priv->soc->mdio_adjust_link(priv, i);
- + spin_unlock_irqrestore(&priv->phy->lock, flags);
- + } else if (priv->phy->phy[i]) {
- + phy_disconnect(priv->phy->phy[i]);
- + } else if (priv->mii_bus && priv->mii_bus->phy_map[i]) {
- + phy_detach(priv->mii_bus->phy_map[i]);
- + }
- +}
- +
- +static void fe_phy_start(struct fe_priv *priv)
- +{
- + unsigned long flags;
- + int i;
- +
- + for (i = 0; i < 8; i++) {
- + if (priv->phy->phy_fixed[i]) {
- + spin_lock_irqsave(&priv->phy->lock, flags);
- + priv->link[i] = 1;
- + if (priv->soc->mdio_adjust_link)
- + priv->soc->mdio_adjust_link(priv, i);
- + spin_unlock_irqrestore(&priv->phy->lock, flags);
- + } else if (priv->phy->phy[i]) {
- + phy_start(priv->phy->phy[i]);
- + }
- + }
- +}
- +
- +static void fe_phy_stop(struct fe_priv *priv)
- +{
- + unsigned long flags;
- + int i;
- +
- + for (i = 0; i < 8; i++)
- + if (priv->phy->phy_fixed[i]) {
- + spin_lock_irqsave(&priv->phy->lock, flags);
- + priv->link[i] = 0;
- + if (priv->soc->mdio_adjust_link)
- + priv->soc->mdio_adjust_link(priv, i);
- + spin_unlock_irqrestore(&priv->phy->lock, flags);
- + } else if (priv->phy->phy[i]) {
- + phy_stop(priv->phy->phy[i]);
- + }
- +}
- +
- +static struct fe_phy phy_ralink = {
- + .connect = fe_phy_connect,
- + .disconnect = fe_phy_disconnect,
- + .start = fe_phy_start,
- + .stop = fe_phy_stop,
- +};
- +
- +int fe_mdio_init(struct fe_priv *priv)
- +{
- + struct device_node *mii_np;
- + int err;
- +
- + if (!priv->soc->mdio_read || !priv->soc->mdio_write)
- + return 0;
- +
- + spin_lock_init(&phy_ralink.lock);
- + priv->phy = &phy_ralink;
- +
- + mii_np = of_get_child_by_name(priv->device->of_node, "mdio-bus");
- + if (!mii_np) {
- + dev_err(priv->device, "no %s child node found", "mdio-bus");
- + return -ENODEV;
- + }
- +
- + if (!of_device_is_available(mii_np)) {
- + err = 0;
- + goto err_put_node;
- + }
- +
- + priv->mii_bus = mdiobus_alloc();
- + if (!priv->mii_bus) {
- + err = -ENOMEM;
- + goto err_put_node;
- + }
- +
- + priv->mii_bus->name = "mdio";
- + priv->mii_bus->read = priv->soc->mdio_read;
- + priv->mii_bus->write = priv->soc->mdio_write;
- + priv->mii_bus->reset = fe_mdio_reset;
- + priv->mii_bus->priv = priv;
- + priv->mii_bus->parent = priv->device;
- +
- + snprintf(priv->mii_bus->id, MII_BUS_ID_SIZE, "%s", mii_np->name);
- + err = of_mdiobus_register(priv->mii_bus, mii_np);
- + if (err)
- + goto err_free_bus;
- +
- + return 0;
- +
- +err_free_bus:
- + kfree(priv->mii_bus);
- +err_put_node:
- + of_node_put(mii_np);
- + priv->mii_bus = NULL;
- + return err;
- +}
- +
- +void fe_mdio_cleanup(struct fe_priv *priv)
- +{
- + if (!priv->mii_bus)
- + return;
- +
- + mdiobus_unregister(priv->mii_bus);
- + of_node_put(priv->mii_bus->dev.of_node);
- + kfree(priv->mii_bus);
- +}
- --- /dev/null
- +++ b/drivers/net/ethernet/mediatek/mdio.h
- @@ -0,0 +1,27 @@
- +/* This program is free software; you can redistribute it and/or modify
- + * it under the terms of the GNU General Public License as published by
- + * the Free Software Foundation; version 2 of the License
- + *
- + * This program is distributed in the hope that it will be useful,
- + * but WITHOUT ANY WARRANTY; without even the implied warranty of
- + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- + * GNU General Public License for more details.
- + *
- + * Copyright (C) 2009-2015 John Crispin <blogic@openwrt.org>
- + * Copyright (C) 2009-2015 Felix Fietkau <nbd@nbd.name>
- + * Copyright (C) 2013-2015 Michael Lee <igvtee@gmail.com>
- + */
- +
- +#ifndef _RALINK_MDIO_H__
- +#define _RALINK_MDIO_H__
- +
- +#ifdef CONFIG_NET_MEDIATEK_MDIO
- +int fe_mdio_init(struct fe_priv *priv);
- +void fe_mdio_cleanup(struct fe_priv *priv);
- +int fe_connect_phy_node(struct fe_priv *priv,
- + struct device_node *phy_node);
- +#else
- +static inline int fe_mdio_init(struct fe_priv *priv) { return 0; }
- +static inline void fe_mdio_cleanup(struct fe_priv *priv) {}
- +#endif
- +#endif
- --- /dev/null
- +++ b/drivers/net/ethernet/mediatek/mtk_eth_soc.c
- @@ -0,0 +1,1587 @@
- +/* This program is free software; you can redistribute it and/or modify
- + * it under the terms of the GNU General Public License as published by
- + * the Free Software Foundation; version 2 of the License
- + *
- + * This program is distributed in the hope that it will be useful,
- + * but WITHOUT ANY WARRANTY; without even the implied warranty of
- + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- + * GNU General Public License for more details.
- + *
- + * Copyright (C) 2009-2015 John Crispin <blogic@openwrt.org>
- + * Copyright (C) 2009-2015 Felix Fietkau <nbd@nbd.name>
- + * Copyright (C) 2013-2015 Michael Lee <igvtee@gmail.com>
- + */
- +
- +#include <linux/module.h>
- +#include <linux/kernel.h>
- +#include <linux/types.h>
- +#include <linux/dma-mapping.h>
- +#include <linux/init.h>
- +#include <linux/skbuff.h>
- +#include <linux/etherdevice.h>
- +#include <linux/ethtool.h>
- +#include <linux/platform_device.h>
- +#include <linux/of_device.h>
- +#include <linux/clk.h>
- +#include <linux/of_net.h>
- +#include <linux/of_mdio.h>
- +#include <linux/if_vlan.h>
- +#include <linux/reset.h>
- +#include <linux/tcp.h>
- +#include <linux/io.h>
- +#include <linux/bug.h>
- +
- +#include <asm/mach-ralink/ralink_regs.h>
- +
- +#include "mtk_eth_soc.h"
- +#include "mdio.h"
- +#include "ethtool.h"
- +
- +#define MAX_RX_LENGTH 1536
- +#define FE_RX_ETH_HLEN (VLAN_ETH_HLEN + VLAN_HLEN + ETH_FCS_LEN)
- +#define FE_RX_HLEN (NET_SKB_PAD + FE_RX_ETH_HLEN + NET_IP_ALIGN)
- +#define DMA_DUMMY_DESC 0xffffffff
- +#define FE_DEFAULT_MSG_ENABLE \
- + (NETIF_MSG_DRV | \
- + NETIF_MSG_PROBE | \
- + NETIF_MSG_LINK | \
- + NETIF_MSG_TIMER | \
- + NETIF_MSG_IFDOWN | \
- + NETIF_MSG_IFUP | \
- + NETIF_MSG_RX_ERR | \
- + NETIF_MSG_TX_ERR)
- +
- +#define TX_DMA_DESP2_DEF (TX_DMA_LS0 | TX_DMA_DONE)
- +#define TX_DMA_DESP4_DEF (TX_DMA_QN(3) | TX_DMA_PN(1))
- +#define NEXT_TX_DESP_IDX(X) (((X) + 1) & (ring->tx_ring_size - 1))
- +#define NEXT_RX_DESP_IDX(X) (((X) + 1) & (ring->rx_ring_size - 1))
- +
- +#define SYSC_REG_RSTCTRL 0x34
- +
- +static int fe_msg_level = -1;
- +module_param_named(msg_level, fe_msg_level, int, 0);
- +MODULE_PARM_DESC(msg_level, "Message level (-1=defaults,0=none,...,16=all)");
- +
- +static const u16 fe_reg_table_default[FE_REG_COUNT] = {
- + [FE_REG_PDMA_GLO_CFG] = FE_PDMA_GLO_CFG,
- + [FE_REG_PDMA_RST_CFG] = FE_PDMA_RST_CFG,
- + [FE_REG_DLY_INT_CFG] = FE_DLY_INT_CFG,
- + [FE_REG_TX_BASE_PTR0] = FE_TX_BASE_PTR0,
- + [FE_REG_TX_MAX_CNT0] = FE_TX_MAX_CNT0,
- + [FE_REG_TX_CTX_IDX0] = FE_TX_CTX_IDX0,
- + [FE_REG_TX_DTX_IDX0] = FE_TX_DTX_IDX0,
- + [FE_REG_RX_BASE_PTR0] = FE_RX_BASE_PTR0,
- + [FE_REG_RX_MAX_CNT0] = FE_RX_MAX_CNT0,
- + [FE_REG_RX_CALC_IDX0] = FE_RX_CALC_IDX0,
- + [FE_REG_RX_DRX_IDX0] = FE_RX_DRX_IDX0,
- + [FE_REG_FE_INT_ENABLE] = FE_FE_INT_ENABLE,
- + [FE_REG_FE_INT_STATUS] = FE_FE_INT_STATUS,
- + [FE_REG_FE_DMA_VID_BASE] = FE_DMA_VID0,
- + [FE_REG_FE_COUNTER_BASE] = FE_GDMA1_TX_GBCNT,
- + [FE_REG_FE_RST_GL] = FE_FE_RST_GL,
- +};
- +
- +static const u16 *fe_reg_table = fe_reg_table_default;
- +
- +struct fe_work_t {
- + int bitnr;
- + void (*action)(struct fe_priv *);
- +};
- +
- +static void __iomem *fe_base;
- +
- +void fe_w32(u32 val, unsigned reg)
- +{
- + __raw_writel(val, fe_base + reg);
- +}
- +
- +u32 fe_r32(unsigned reg)
- +{
- + return __raw_readl(fe_base + reg);
- +}
- +
- +void fe_reg_w32(u32 val, enum fe_reg reg)
- +{
- + fe_w32(val, fe_reg_table[reg]);
- +}
- +
- +u32 fe_reg_r32(enum fe_reg reg)
- +{
- + return fe_r32(fe_reg_table[reg]);
- +}
- +
- +void fe_reset(u32 reset_bits)
- +{
- + u32 t;
- +
- + t = rt_sysc_r32(SYSC_REG_RSTCTRL);
- + t |= reset_bits;
- + rt_sysc_w32(t, SYSC_REG_RSTCTRL);
- + usleep_range(10, 20);
- +
- + t &= ~reset_bits;
- + rt_sysc_w32(t, SYSC_REG_RSTCTRL);
- + usleep_range(10, 20);
- +}
- +
- +static inline void fe_int_disable(u32 mask)
- +{
- + fe_reg_w32(fe_reg_r32(FE_REG_FE_INT_ENABLE) & ~mask,
- + FE_REG_FE_INT_ENABLE);
- + /* flush write */
- + fe_reg_r32(FE_REG_FE_INT_ENABLE);
- +}
- +
- +static inline void fe_int_enable(u32 mask)
- +{
- + fe_reg_w32(fe_reg_r32(FE_REG_FE_INT_ENABLE) | mask,
- + FE_REG_FE_INT_ENABLE);
- + /* flush write */
- + fe_reg_r32(FE_REG_FE_INT_ENABLE);
- +}
- +
- +static inline void fe_hw_set_macaddr(struct fe_priv *priv, unsigned char *mac)
- +{
- + unsigned long flags;
- +
- + spin_lock_irqsave(&priv->page_lock, flags);
- + fe_w32((mac[0] << 8) | mac[1], FE_GDMA1_MAC_ADRH);
- + fe_w32((mac[2] << 24) | (mac[3] << 16) | (mac[4] << 8) | mac[5],
- + FE_GDMA1_MAC_ADRL);
- + spin_unlock_irqrestore(&priv->page_lock, flags);
- +}
- +
- +static int fe_set_mac_address(struct net_device *dev, void *p)
- +{
- + int ret = eth_mac_addr(dev, p);
- +
- + if (!ret) {
- + struct fe_priv *priv = netdev_priv(dev);
- +
- + if (priv->soc->set_mac)
- + priv->soc->set_mac(priv, dev->dev_addr);
- + else
- + fe_hw_set_macaddr(priv, p);
- + }
- +
- + return ret;
- +}
- +
- +static inline int fe_max_frag_size(int mtu)
- +{
- + /* make sure buf_size will be at least MAX_RX_LENGTH */
- + if (mtu + FE_RX_ETH_HLEN < MAX_RX_LENGTH)
- + mtu = MAX_RX_LENGTH - FE_RX_ETH_HLEN;
- +
- + return SKB_DATA_ALIGN(FE_RX_HLEN + mtu) +
- + SKB_DATA_ALIGN(sizeof(struct skb_shared_info));
- +}
- +
- +static inline int fe_max_buf_size(int frag_size)
- +{
- + int buf_size = frag_size - NET_SKB_PAD - NET_IP_ALIGN -
- + SKB_DATA_ALIGN(sizeof(struct skb_shared_info));
- +
- + BUG_ON(buf_size < MAX_RX_LENGTH);
- + return buf_size;
- +}
- +
- +static inline void fe_get_rxd(struct fe_rx_dma *rxd, struct fe_rx_dma *dma_rxd)
- +{
- + rxd->rxd1 = dma_rxd->rxd1;
- + rxd->rxd2 = dma_rxd->rxd2;
- + rxd->rxd3 = dma_rxd->rxd3;
- + rxd->rxd4 = dma_rxd->rxd4;
- +}
- +
- +static inline void fe_set_txd(struct fe_tx_dma *txd, struct fe_tx_dma *dma_txd)
- +{
- + dma_txd->txd1 = txd->txd1;
- + dma_txd->txd3 = txd->txd3;
- + dma_txd->txd4 = txd->txd4;
- + /* clean dma done flag last */
- + dma_txd->txd2 = txd->txd2;
- +}
- +
- +static void fe_clean_rx(struct fe_priv *priv)
- +{
- + int i;
- + struct fe_rx_ring *ring = &priv->rx_ring;
- +
- + if (ring->rx_data) {
- + for (i = 0; i < ring->rx_ring_size; i++)
- + if (ring->rx_data[i]) {
- + if (ring->rx_dma && ring->rx_dma[i].rxd1)
- + dma_unmap_single(&priv->netdev->dev,
- + ring->rx_dma[i].rxd1,
- + ring->rx_buf_size,
- + DMA_FROM_DEVICE);
- + put_page(virt_to_head_page(ring->rx_data[i]));
- + }
- +
- + kfree(ring->rx_data);
- + ring->rx_data = NULL;
- + }
- +
- + if (ring->rx_dma) {
- + dma_free_coherent(&priv->netdev->dev,
- + ring->rx_ring_size * sizeof(*ring->rx_dma),
- + ring->rx_dma,
- + ring->rx_phys);
- + ring->rx_dma = NULL;
- + }
- +}
- +
- +static int fe_alloc_rx(struct fe_priv *priv)
- +{
- + struct net_device *netdev = priv->netdev;
- + struct fe_rx_ring *ring = &priv->rx_ring;
- + int i, pad;
- +
- + ring->rx_data = kcalloc(ring->rx_ring_size, sizeof(*ring->rx_data),
- + GFP_KERNEL);
- + if (!ring->rx_data)
- + goto no_rx_mem;
- +
- + for (i = 0; i < ring->rx_ring_size; i++) {
- + ring->rx_data[i] = netdev_alloc_frag(ring->frag_size);
- + if (!ring->rx_data[i])
- + goto no_rx_mem;
- + }
- +
- + ring->rx_dma = dma_alloc_coherent(&netdev->dev,
- + ring->rx_ring_size * sizeof(*ring->rx_dma),
- + &ring->rx_phys,
- + GFP_ATOMIC | __GFP_ZERO);
- + if (!ring->rx_dma)
- + goto no_rx_mem;
- +
- + if (priv->flags & FE_FLAG_RX_2B_OFFSET)
- + pad = 0;
- + else
- + pad = NET_IP_ALIGN;
- + for (i = 0; i < ring->rx_ring_size; i++) {
- + dma_addr_t dma_addr = dma_map_single(&netdev->dev,
- + ring->rx_data[i] + NET_SKB_PAD + pad,
- + ring->rx_buf_size,
- + DMA_FROM_DEVICE);
- + if (unlikely(dma_mapping_error(&netdev->dev, dma_addr)))
- + goto no_rx_mem;
- + ring->rx_dma[i].rxd1 = (unsigned int)dma_addr;
- +
- + if (priv->flags & FE_FLAG_RX_SG_DMA)
- + ring->rx_dma[i].rxd2 = RX_DMA_PLEN0(ring->rx_buf_size);
- + else
- + ring->rx_dma[i].rxd2 = RX_DMA_LSO;
- + }
- + ring->rx_calc_idx = ring->rx_ring_size - 1;
- + /* make sure that all changes to the dma ring are flushed before we
- + * continue
- + */
- + wmb();
- +
- + fe_reg_w32(ring->rx_phys, FE_REG_RX_BASE_PTR0);
- + fe_reg_w32(ring->rx_ring_size, FE_REG_RX_MAX_CNT0);
- + fe_reg_w32(ring->rx_calc_idx, FE_REG_RX_CALC_IDX0);
- + fe_reg_w32(FE_PST_DRX_IDX0, FE_REG_PDMA_RST_CFG);
- +
- + return 0;
- +
- +no_rx_mem:
- + return -ENOMEM;
- +}
- +
- +static void fe_txd_unmap(struct device *dev, struct fe_tx_buf *tx_buf)
- +{
- + if (tx_buf->flags & FE_TX_FLAGS_SINGLE0) {
- + dma_unmap_single(dev,
- + dma_unmap_addr(tx_buf, dma_addr0),
- + dma_unmap_len(tx_buf, dma_len0),
- + DMA_TO_DEVICE);
- + } else if (tx_buf->flags & FE_TX_FLAGS_PAGE0) {
- + dma_unmap_page(dev,
- + dma_unmap_addr(tx_buf, dma_addr0),
- + dma_unmap_len(tx_buf, dma_len0),
- + DMA_TO_DEVICE);
- + }
- + if (tx_buf->flags & FE_TX_FLAGS_PAGE1)
- + dma_unmap_page(dev,
- + dma_unmap_addr(tx_buf, dma_addr1),
- + dma_unmap_len(tx_buf, dma_len1),
- + DMA_TO_DEVICE);
- +
- + tx_buf->flags = 0;
- + if (tx_buf->skb && (tx_buf->skb != (struct sk_buff *)DMA_DUMMY_DESC))
- + dev_kfree_skb_any(tx_buf->skb);
- + tx_buf->skb = NULL;
- +}
- +
- +static void fe_clean_tx(struct fe_priv *priv)
- +{
- + int i;
- + struct device *dev = &priv->netdev->dev;
- + struct fe_tx_ring *ring = &priv->tx_ring;
- +
- + if (ring->tx_buf) {
- + for (i = 0; i < ring->tx_ring_size; i++)
- + fe_txd_unmap(dev, &ring->tx_buf[i]);
- + kfree(ring->tx_buf);
- + ring->tx_buf = NULL;
- + }
- +
- + if (ring->tx_dma) {
- + dma_free_coherent(dev,
- + ring->tx_ring_size * sizeof(*ring->tx_dma),
- + ring->tx_dma,
- + ring->tx_phys);
- + ring->tx_dma = NULL;
- + }
- +
- + netdev_reset_queue(priv->netdev);
- +}
- +
- +static int fe_alloc_tx(struct fe_priv *priv)
- +{
- + int i;
- + struct fe_tx_ring *ring = &priv->tx_ring;
- +
- + ring->tx_free_idx = 0;
- + ring->tx_next_idx = 0;
- + ring->tx_thresh = max((unsigned long)ring->tx_ring_size >> 2,
- + MAX_SKB_FRAGS);
- +
- + ring->tx_buf = kcalloc(ring->tx_ring_size, sizeof(*ring->tx_buf),
- + GFP_KERNEL);
- + if (!ring->tx_buf)
- + goto no_tx_mem;
- +
- + ring->tx_dma = dma_alloc_coherent(&priv->netdev->dev,
- + ring->tx_ring_size * sizeof(*ring->tx_dma),
- + &ring->tx_phys,
- + GFP_ATOMIC | __GFP_ZERO);
- + if (!ring->tx_dma)
- + goto no_tx_mem;
- +
- + for (i = 0; i < ring->tx_ring_size; i++) {
- + if (priv->soc->tx_dma)
- + priv->soc->tx_dma(&ring->tx_dma[i]);
- + ring->tx_dma[i].txd2 = TX_DMA_DESP2_DEF;
- + }
- + /* make sure that all changes to the dma ring are flushed before we
- + * continue
- + */
- + wmb();
- +
- + fe_reg_w32(ring->tx_phys, FE_REG_TX_BASE_PTR0);
- + fe_reg_w32(ring->tx_ring_size, FE_REG_TX_MAX_CNT0);
- + fe_reg_w32(0, FE_REG_TX_CTX_IDX0);
- + fe_reg_w32(FE_PST_DTX_IDX0, FE_REG_PDMA_RST_CFG);
- +
- + return 0;
- +
- +no_tx_mem:
- + return -ENOMEM;
- +}
- +
- +static int fe_init_dma(struct fe_priv *priv)
- +{
- + int err;
- +
- + err = fe_alloc_tx(priv);
- + if (err)
- + return err;
- +
- + err = fe_alloc_rx(priv);
- + if (err)
- + return err;
- +
- + return 0;
- +}
- +
- +static void fe_free_dma(struct fe_priv *priv)
- +{
- + fe_clean_tx(priv);
- + fe_clean_rx(priv);
- +}
- +
- +void fe_stats_update(struct fe_priv *priv)
- +{
- + struct fe_hw_stats *hwstats = priv->hw_stats;
- + unsigned int base = fe_reg_table[FE_REG_FE_COUNTER_BASE];
- + u64 stats;
- +
- + u64_stats_update_begin(&hwstats->syncp);
- +
- + if (IS_ENABLED(CONFIG_SOC_MT7621)) {
- + hwstats->rx_bytes += fe_r32(base);
- + stats = fe_r32(base + 0x04);
- + if (stats)
- + hwstats->rx_bytes += (stats << 32);
- + hwstats->rx_packets += fe_r32(base + 0x08);
- + hwstats->rx_overflow += fe_r32(base + 0x10);
- + hwstats->rx_fcs_errors += fe_r32(base + 0x14);
- + hwstats->rx_short_errors += fe_r32(base + 0x18);
- + hwstats->rx_long_errors += fe_r32(base + 0x1c);
- + hwstats->rx_checksum_errors += fe_r32(base + 0x20);
- + hwstats->rx_flow_control_packets += fe_r32(base + 0x24);
- + hwstats->tx_skip += fe_r32(base + 0x28);
- + hwstats->tx_collisions += fe_r32(base + 0x2c);
- + hwstats->tx_bytes += fe_r32(base + 0x30);
- + stats = fe_r32(base + 0x34);
- + if (stats)
- + hwstats->tx_bytes += (stats << 32);
- + hwstats->tx_packets += fe_r32(base + 0x38);
- + } else {
- + hwstats->tx_bytes += fe_r32(base);
- + hwstats->tx_packets += fe_r32(base + 0x04);
- + hwstats->tx_skip += fe_r32(base + 0x08);
- + hwstats->tx_collisions += fe_r32(base + 0x0c);
- + hwstats->rx_bytes += fe_r32(base + 0x20);
- + hwstats->rx_packets += fe_r32(base + 0x24);
- + hwstats->rx_overflow += fe_r32(base + 0x28);
- + hwstats->rx_fcs_errors += fe_r32(base + 0x2c);
- + hwstats->rx_short_errors += fe_r32(base + 0x30);
- + hwstats->rx_long_errors += fe_r32(base + 0x34);
- + hwstats->rx_checksum_errors += fe_r32(base + 0x38);
- + hwstats->rx_flow_control_packets += fe_r32(base + 0x3c);
- + }
- +
- + u64_stats_update_end(&hwstats->syncp);
- +}
- +
- +static struct rtnl_link_stats64 *fe_get_stats64(struct net_device *dev,
- + struct rtnl_link_stats64 *storage)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- + struct fe_hw_stats *hwstats = priv->hw_stats;
- + unsigned int base = fe_reg_table[FE_REG_FE_COUNTER_BASE];
- + unsigned int start;
- +
- + if (!base) {
- + netdev_stats_to_stats64(storage, &dev->stats);
- + return storage;
- + }
- +
- + if (netif_running(dev) && netif_device_present(dev)) {
- + if (spin_trylock(&hwstats->stats_lock)) {
- + fe_stats_update(priv);
- + spin_unlock(&hwstats->stats_lock);
- + }
- + }
- +
- + do {
- + start = u64_stats_fetch_begin_irq(&hwstats->syncp);
- + storage->rx_packets = hwstats->rx_packets;
- + storage->tx_packets = hwstats->tx_packets;
- + storage->rx_bytes = hwstats->rx_bytes;
- + storage->tx_bytes = hwstats->tx_bytes;
- + storage->collisions = hwstats->tx_collisions;
- + storage->rx_length_errors = hwstats->rx_short_errors +
- + hwstats->rx_long_errors;
- + storage->rx_over_errors = hwstats->rx_overflow;
- + storage->rx_crc_errors = hwstats->rx_fcs_errors;
- + storage->rx_errors = hwstats->rx_checksum_errors;
- + storage->tx_aborted_errors = hwstats->tx_skip;
- + } while (u64_stats_fetch_retry_irq(&hwstats->syncp, start));
- +
- + storage->tx_errors = priv->netdev->stats.tx_errors;
- + storage->rx_dropped = priv->netdev->stats.rx_dropped;
- + storage->tx_dropped = priv->netdev->stats.tx_dropped;
- +
- + return storage;
- +}
- +
- +static int fe_vlan_rx_add_vid(struct net_device *dev,
- + __be16 proto, u16 vid)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- + u32 idx = (vid & 0xf);
- + u32 vlan_cfg;
- +
- + if (!((fe_reg_table[FE_REG_FE_DMA_VID_BASE]) &&
- + (dev->features & NETIF_F_HW_VLAN_CTAG_TX)))
- + return 0;
- +
- + if (test_bit(idx, &priv->vlan_map)) {
- + netdev_warn(dev, "disable tx vlan offload\n");
- + dev->wanted_features &= ~NETIF_F_HW_VLAN_CTAG_TX;
- + netdev_update_features(dev);
- + } else {
- + vlan_cfg = fe_r32(fe_reg_table[FE_REG_FE_DMA_VID_BASE] +
- + ((idx >> 1) << 2));
- + if (idx & 0x1) {
- + vlan_cfg &= 0xffff;
- + vlan_cfg |= (vid << 16);
- + } else {
- + vlan_cfg &= 0xffff0000;
- + vlan_cfg |= vid;
- + }
- + fe_w32(vlan_cfg, fe_reg_table[FE_REG_FE_DMA_VID_BASE] +
- + ((idx >> 1) << 2));
- + set_bit(idx, &priv->vlan_map);
- + }
- +
- + return 0;
- +}
- +
- +static int fe_vlan_rx_kill_vid(struct net_device *dev,
- + __be16 proto, u16 vid)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- + u32 idx = (vid & 0xf);
- +
- + if (!((fe_reg_table[FE_REG_FE_DMA_VID_BASE]) &&
- + (dev->features & NETIF_F_HW_VLAN_CTAG_TX)))
- + return 0;
- +
- + clear_bit(idx, &priv->vlan_map);
- +
- + return 0;
- +}
- +
- +static inline u32 fe_empty_txd(struct fe_tx_ring *ring)
- +{
- + barrier();
- + return (u32)(ring->tx_ring_size -
- + ((ring->tx_next_idx - ring->tx_free_idx) &
- + (ring->tx_ring_size - 1)));
- +}
- +
- +static int fe_tx_map_dma(struct sk_buff *skb, struct net_device *dev,
- + int tx_num, struct fe_tx_ring *ring)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- + struct skb_frag_struct *frag;
- + struct fe_tx_dma txd, *ptxd;
- + struct fe_tx_buf *tx_buf;
- + dma_addr_t mapped_addr;
- + unsigned int nr_frags;
- + u32 def_txd4;
- + int i, j, k, frag_size, frag_map_size, offset;
- +
- + tx_buf = &ring->tx_buf[ring->tx_next_idx];
- + memset(tx_buf, 0, sizeof(*tx_buf));
- + memset(&txd, 0, sizeof(txd));
- + nr_frags = skb_shinfo(skb)->nr_frags;
- +
- + /* init tx descriptor */
- + if (priv->soc->tx_dma)
- + priv->soc->tx_dma(&txd);
- + else
- + txd.txd4 = TX_DMA_DESP4_DEF;
- + def_txd4 = txd.txd4;
- +
- + /* TX Checksum offload */
- + if (skb->ip_summed == CHECKSUM_PARTIAL)
- + txd.txd4 |= TX_DMA_CHKSUM;
- +
- + /* VLAN header offload */
- + if (skb_vlan_tag_present(skb)) {
- + u16 tag = skb_vlan_tag_get(skb);
- +
- + if (IS_ENABLED(CONFIG_SOC_MT7621))
- + txd.txd4 |= TX_DMA_INS_VLAN_MT7621 | tag;
- + else
- + txd.txd4 |= TX_DMA_INS_VLAN |
- + ((tag >> VLAN_PRIO_SHIFT) << 4) |
- + (tag & 0xF);
- + }
- +
- + /* TSO: fill MSS info in tcp checksum field */
- + if (skb_is_gso(skb)) {
- + if (skb_cow_head(skb, 0)) {
- + netif_warn(priv, tx_err, dev,
- + "GSO expand head fail.\n");
- + goto err_out;
- + }
- + if (skb_shinfo(skb)->gso_type &
- + (SKB_GSO_TCPV4 | SKB_GSO_TCPV6)) {
- + txd.txd4 |= TX_DMA_TSO;
- + tcp_hdr(skb)->check = htons(skb_shinfo(skb)->gso_size);
- + }
- + }
- +
- + mapped_addr = dma_map_single(&dev->dev, skb->data,
- + skb_headlen(skb), DMA_TO_DEVICE);
- + if (unlikely(dma_mapping_error(&dev->dev, mapped_addr)))
- + goto err_out;
- + txd.txd1 = mapped_addr;
- + txd.txd2 = TX_DMA_PLEN0(skb_headlen(skb));
- +
- + tx_buf->flags |= FE_TX_FLAGS_SINGLE0;
- + dma_unmap_addr_set(tx_buf, dma_addr0, mapped_addr);
- + dma_unmap_len_set(tx_buf, dma_len0, skb_headlen(skb));
- +
- + /* TX SG offload */
- + j = ring->tx_next_idx;
- + k = 0;
- + for (i = 0; i < nr_frags; i++) {
- + offset = 0;
- + frag = &skb_shinfo(skb)->frags[i];
- + frag_size = skb_frag_size(frag);
- +
- + while (frag_size > 0) {
- + frag_map_size = min(frag_size, TX_DMA_BUF_LEN);
- + mapped_addr = skb_frag_dma_map(&dev->dev, frag, offset,
- + frag_map_size,
- + DMA_TO_DEVICE);
- + if (unlikely(dma_mapping_error(&dev->dev, mapped_addr)))
- + goto err_dma;
- +
- + if (k & 0x1) {
- + j = NEXT_TX_DESP_IDX(j);
- + txd.txd1 = mapped_addr;
- + txd.txd2 = TX_DMA_PLEN0(frag_map_size);
- + txd.txd4 = def_txd4;
- +
- + tx_buf = &ring->tx_buf[j];
- + memset(tx_buf, 0, sizeof(*tx_buf));
- +
- + tx_buf->flags |= FE_TX_FLAGS_PAGE0;
- + dma_unmap_addr_set(tx_buf, dma_addr0,
- + mapped_addr);
- + dma_unmap_len_set(tx_buf, dma_len0,
- + frag_map_size);
- + } else {
- + txd.txd3 = mapped_addr;
- + txd.txd2 |= TX_DMA_PLEN1(frag_map_size);
- +
- + tx_buf->skb = (struct sk_buff *)DMA_DUMMY_DESC;
- + tx_buf->flags |= FE_TX_FLAGS_PAGE1;
- + dma_unmap_addr_set(tx_buf, dma_addr1,
- + mapped_addr);
- + dma_unmap_len_set(tx_buf, dma_len1,
- + frag_map_size);
- +
- + if (!((i == (nr_frags - 1)) &&
- + (frag_map_size == frag_size))) {
- + fe_set_txd(&txd, &ring->tx_dma[j]);
- + memset(&txd, 0, sizeof(txd));
- + }
- + }
- + frag_size -= frag_map_size;
- + offset += frag_map_size;
- + k++;
- + }
- + }
- +
- + /* set last segment */
- + if (k & 0x1)
- + txd.txd2 |= TX_DMA_LS1;
- + else
- + txd.txd2 |= TX_DMA_LS0;
- + fe_set_txd(&txd, &ring->tx_dma[j]);
- +
- + /* store skb to cleanup */
- + tx_buf->skb = skb;
- +
- + netdev_sent_queue(dev, skb->len);
- + skb_tx_timestamp(skb);
- +
- + ring->tx_next_idx = NEXT_TX_DESP_IDX(j);
- + /* make sure that all changes to the dma ring are flushed before we
- + * continue
- + */
- + wmb();
- + if (unlikely(fe_empty_txd(ring) <= ring->tx_thresh)) {
- + netif_stop_queue(dev);
- + smp_mb();
- + if (unlikely(fe_empty_txd(ring) > ring->tx_thresh))
- + netif_wake_queue(dev);
- + }
- +
- + if (netif_xmit_stopped(netdev_get_tx_queue(dev, 0)) || !skb->xmit_more)
- + fe_reg_w32(ring->tx_next_idx, FE_REG_TX_CTX_IDX0);
- +
- + return 0;
- +
- +err_dma:
- + j = ring->tx_next_idx;
- + for (i = 0; i < tx_num; i++) {
- + ptxd = &ring->tx_dma[j];
- + tx_buf = &ring->tx_buf[j];
- +
- + /* unmap dma */
- + fe_txd_unmap(&dev->dev, tx_buf);
- +
- + ptxd->txd2 = TX_DMA_DESP2_DEF;
- + j = NEXT_TX_DESP_IDX(j);
- + }
- + /* make sure that all changes to the dma ring are flushed before we
- + * continue
- + */
- + wmb();
- +
- +err_out:
- + return -1;
- +}
- +
- +static inline int fe_skb_padto(struct sk_buff *skb, struct fe_priv *priv)
- +{
- + unsigned int len;
- + int ret;
- +
- + ret = 0;
- + if (unlikely(skb->len < VLAN_ETH_ZLEN)) {
- + if ((priv->flags & FE_FLAG_PADDING_64B) &&
- + !(priv->flags & FE_FLAG_PADDING_BUG))
- + return ret;
- +
- + if (skb_vlan_tag_present(skb))
- + len = ETH_ZLEN;
- + else if (skb->protocol == cpu_to_be16(ETH_P_8021Q))
- + len = VLAN_ETH_ZLEN;
- + else if (!(priv->flags & FE_FLAG_PADDING_64B))
- + len = ETH_ZLEN;
- + else
- + return ret;
- +
- + if (skb->len < len) {
- + ret = skb_pad(skb, len - skb->len);
- + if (ret < 0)
- + return ret;
- + skb->len = len;
- + skb_set_tail_pointer(skb, len);
- + }
- + }
- +
- + return ret;
- +}
- +
- +static inline int fe_cal_txd_req(struct sk_buff *skb)
- +{
- + int i, nfrags;
- + struct skb_frag_struct *frag;
- +
- + nfrags = 1;
- + if (skb_is_gso(skb)) {
- + for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
- + frag = &skb_shinfo(skb)->frags[i];
- + nfrags += DIV_ROUND_UP(frag->size, TX_DMA_BUF_LEN);
- + }
- + } else {
- + nfrags += skb_shinfo(skb)->nr_frags;
- + }
- +
- + return DIV_ROUND_UP(nfrags, 2);
- +}
- +
- +static int fe_start_xmit(struct sk_buff *skb, struct net_device *dev)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- + struct fe_tx_ring *ring = &priv->tx_ring;
- + struct net_device_stats *stats = &dev->stats;
- + int tx_num;
- + int len = skb->len;
- +
- + if (fe_skb_padto(skb, priv)) {
- + netif_warn(priv, tx_err, dev, "tx padding failed!\n");
- + return NETDEV_TX_OK;
- + }
- +
- + tx_num = fe_cal_txd_req(skb);
- + if (unlikely(fe_empty_txd(ring) <= tx_num)) {
- + netif_stop_queue(dev);
- + netif_err(priv, tx_queued, dev,
- + "Tx Ring full when queue awake!\n");
- + return NETDEV_TX_BUSY;
- + }
- +
- + if (fe_tx_map_dma(skb, dev, tx_num, ring) < 0) {
- + stats->tx_dropped++;
- + } else {
- + stats->tx_packets++;
- + stats->tx_bytes += len;
- + }
- +
- + return NETDEV_TX_OK;
- +}
- +
- +static int fe_poll_rx(struct napi_struct *napi, int budget,
- + struct fe_priv *priv, u32 rx_intr)
- +{
- + struct net_device *netdev = priv->netdev;
- + struct net_device_stats *stats = &netdev->stats;
- + struct fe_soc_data *soc = priv->soc;
- + struct fe_rx_ring *ring = &priv->rx_ring;
- + int idx = ring->rx_calc_idx;
- + u32 checksum_bit;
- + struct sk_buff *skb;
- + u8 *data, *new_data;
- + struct fe_rx_dma *rxd, trxd;
- + int done = 0, pad;
- +
- + if (netdev->features & NETIF_F_RXCSUM)
- + checksum_bit = soc->checksum_bit;
- + else
- + checksum_bit = 0;
- +
- + if (priv->flags & FE_FLAG_RX_2B_OFFSET)
- + pad = 0;
- + else
- + pad = NET_IP_ALIGN;
- +
- + while (done < budget) {
- + unsigned int pktlen;
- + dma_addr_t dma_addr;
- +
- + idx = NEXT_RX_DESP_IDX(idx);
- + rxd = &ring->rx_dma[idx];
- + data = ring->rx_data[idx];
- +
- + fe_get_rxd(&trxd, rxd);
- + if (!(trxd.rxd2 & RX_DMA_DONE))
- + break;
- +
- + /* alloc new buffer */
- + new_data = netdev_alloc_frag(ring->frag_size);
- + if (unlikely(!new_data)) {
- + stats->rx_dropped++;
- + goto release_desc;
- + }
- + dma_addr = dma_map_single(&netdev->dev,
- + new_data + NET_SKB_PAD + pad,
- + ring->rx_buf_size,
- + DMA_FROM_DEVICE);
- + if (unlikely(dma_mapping_error(&netdev->dev, dma_addr))) {
- + put_page(virt_to_head_page(new_data));
- + goto release_desc;
- + }
- +
- + /* receive data */
- + skb = build_skb(data, ring->frag_size);
- + if (unlikely(!skb)) {
- + put_page(virt_to_head_page(new_data));
- + goto release_desc;
- + }
- + skb_reserve(skb, NET_SKB_PAD + NET_IP_ALIGN);
- +
- + dma_unmap_single(&netdev->dev, trxd.rxd1,
- + ring->rx_buf_size, DMA_FROM_DEVICE);
- + pktlen = RX_DMA_GET_PLEN0(trxd.rxd2);
- + skb->dev = netdev;
- + skb_put(skb, pktlen);
- + if (trxd.rxd4 & checksum_bit)
- + skb->ip_summed = CHECKSUM_UNNECESSARY;
- + else
- + skb_checksum_none_assert(skb);
- + skb->protocol = eth_type_trans(skb, netdev);
- +
- + stats->rx_packets++;
- + stats->rx_bytes += pktlen;
- +
- + napi_gro_receive(napi, skb);
- +
- + ring->rx_data[idx] = new_data;
- + rxd->rxd1 = (unsigned int)dma_addr;
- +
- +release_desc:
- + if (priv->flags & FE_FLAG_RX_SG_DMA)
- + rxd->rxd2 = RX_DMA_PLEN0(ring->rx_buf_size);
- + else
- + rxd->rxd2 = RX_DMA_LSO;
- +
- + ring->rx_calc_idx = idx;
- + /* make sure that all changes to the dma ring are flushed before
- + * we continue
- + */
- + wmb();
- + fe_reg_w32(ring->rx_calc_idx, FE_REG_RX_CALC_IDX0);
- + done++;
- + }
- +
- + if (done < budget)
- + fe_reg_w32(rx_intr, FE_REG_FE_INT_STATUS);
- +
- + return done;
- +}
- +
- +static int fe_poll_tx(struct fe_priv *priv, int budget, u32 tx_intr,
- + int *tx_again)
- +{
- + struct net_device *netdev = priv->netdev;
- + struct device *dev = &netdev->dev;
- + unsigned int bytes_compl = 0;
- + struct sk_buff *skb;
- + struct fe_tx_buf *tx_buf;
- + int done = 0;
- + u32 idx, hwidx;
- + struct fe_tx_ring *ring = &priv->tx_ring;
- +
- + idx = ring->tx_free_idx;
- + hwidx = fe_reg_r32(FE_REG_TX_DTX_IDX0);
- +
- + while ((idx != hwidx) && budget) {
- + tx_buf = &ring->tx_buf[idx];
- + skb = tx_buf->skb;
- +
- + if (!skb)
- + break;
- +
- + if (skb != (struct sk_buff *)DMA_DUMMY_DESC) {
- + bytes_compl += skb->len;
- + done++;
- + budget--;
- + }
- + fe_txd_unmap(dev, tx_buf);
- + idx = NEXT_TX_DESP_IDX(idx);
- + }
- + ring->tx_free_idx = idx;
- +
- + if (idx == hwidx) {
- + /* read hw index again make sure no new tx packet */
- + hwidx = fe_reg_r32(FE_REG_TX_DTX_IDX0);
- + if (idx == hwidx)
- + fe_reg_w32(tx_intr, FE_REG_FE_INT_STATUS);
- + else
- + *tx_again = 1;
- + } else {
- + *tx_again = 1;
- + }
- +
- + if (done) {
- + netdev_completed_queue(netdev, done, bytes_compl);
- + smp_mb();
- + if (unlikely(netif_queue_stopped(netdev) &&
- + (fe_empty_txd(ring) > ring->tx_thresh)))
- + netif_wake_queue(netdev);
- + }
- +
- + return done;
- +}
- +
- +static int fe_poll(struct napi_struct *napi, int budget)
- +{
- + struct fe_priv *priv = container_of(napi, struct fe_priv, rx_napi);
- + struct fe_hw_stats *hwstat = priv->hw_stats;
- + int tx_done, rx_done, tx_again;
- + u32 status, fe_status, status_reg, mask;
- + u32 tx_intr, rx_intr, status_intr;
- +
- + status = fe_reg_r32(FE_REG_FE_INT_STATUS);
- + fe_status = status;
- + tx_intr = priv->soc->tx_int;
- + rx_intr = priv->soc->rx_int;
- + status_intr = priv->soc->status_int;
- + tx_done = 0;
- + rx_done = 0;
- + tx_again = 0;
- +
- + if (fe_reg_table[FE_REG_FE_INT_STATUS2]) {
- + fe_status = fe_reg_r32(FE_REG_FE_INT_STATUS2);
- + status_reg = FE_REG_FE_INT_STATUS2;
- + } else {
- + status_reg = FE_REG_FE_INT_STATUS;
- + }
- +
- + if (status & tx_intr)
- + tx_done = fe_poll_tx(priv, budget, tx_intr, &tx_again);
- +
- + if (status & rx_intr)
- + rx_done = fe_poll_rx(napi, budget, priv, rx_intr);
- +
- + if (unlikely(fe_status & status_intr)) {
- + if (hwstat && spin_trylock(&hwstat->stats_lock)) {
- + fe_stats_update(priv);
- + spin_unlock(&hwstat->stats_lock);
- + }
- + fe_reg_w32(status_intr, status_reg);
- + }
- +
- + if (unlikely(netif_msg_intr(priv))) {
- + mask = fe_reg_r32(FE_REG_FE_INT_ENABLE);
- + netdev_info(priv->netdev,
- + "done tx %d, rx %d, intr 0x%08x/0x%x\n",
- + tx_done, rx_done, status, mask);
- + }
- +
- + if (!tx_again && (rx_done < budget)) {
- + status = fe_reg_r32(FE_REG_FE_INT_STATUS);
- + if (status & (tx_intr | rx_intr)) {
- + /* let napi poll again */
- + rx_done = budget;
- + goto poll_again;
- + }
- +
- + napi_complete(napi);
- + fe_int_enable(tx_intr | rx_intr);
- + } else {
- + rx_done = budget;
- + }
- +
- +poll_again:
- + return rx_done;
- +}
- +
- +static void fe_tx_timeout(struct net_device *dev)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- + struct fe_tx_ring *ring = &priv->tx_ring;
- +
- + priv->netdev->stats.tx_errors++;
- + netif_err(priv, tx_err, dev,
- + "transmit timed out\n");
- + netif_info(priv, drv, dev, "dma_cfg:%08x\n",
- + fe_reg_r32(FE_REG_PDMA_GLO_CFG));
- + netif_info(priv, drv, dev, "tx_ring=%d, "
- + "base=%08x, max=%u, ctx=%u, dtx=%u, fdx=%hu, next=%hu\n",
- + 0, fe_reg_r32(FE_REG_TX_BASE_PTR0),
- + fe_reg_r32(FE_REG_TX_MAX_CNT0),
- + fe_reg_r32(FE_REG_TX_CTX_IDX0),
- + fe_reg_r32(FE_REG_TX_DTX_IDX0),
- + ring->tx_free_idx,
- + ring->tx_next_idx);
- + netif_info(priv, drv, dev,
- + "rx_ring=%d, base=%08x, max=%u, calc=%u, drx=%u\n",
- + 0, fe_reg_r32(FE_REG_RX_BASE_PTR0),
- + fe_reg_r32(FE_REG_RX_MAX_CNT0),
- + fe_reg_r32(FE_REG_RX_CALC_IDX0),
- + fe_reg_r32(FE_REG_RX_DRX_IDX0));
- +
- + if (!test_and_set_bit(FE_FLAG_RESET_PENDING, priv->pending_flags))
- + schedule_work(&priv->pending_work);
- +}
- +
- +static irqreturn_t fe_handle_irq(int irq, void *dev)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- + u32 status, int_mask;
- +
- + status = fe_reg_r32(FE_REG_FE_INT_STATUS);
- +
- + if (unlikely(!status))
- + return IRQ_NONE;
- +
- + int_mask = (priv->soc->rx_int | priv->soc->tx_int);
- + if (likely(status & int_mask)) {
- + if (likely(napi_schedule_prep(&priv->rx_napi))) {
- + fe_int_disable(int_mask);
- + __napi_schedule(&priv->rx_napi);
- + }
- + } else {
- + fe_reg_w32(status, FE_REG_FE_INT_STATUS);
- + }
- +
- + return IRQ_HANDLED;
- +}
- +
- +#ifdef CONFIG_NET_POLL_CONTROLLER
- +static void fe_poll_controller(struct net_device *dev)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- + u32 int_mask = priv->soc->tx_int | priv->soc->rx_int;
- +
- + fe_int_disable(int_mask);
- + fe_handle_irq(dev->irq, dev);
- + fe_int_enable(int_mask);
- +}
- +#endif
- +
- +int fe_set_clock_cycle(struct fe_priv *priv)
- +{
- + unsigned long sysclk = priv->sysclk;
- +
- + sysclk /= FE_US_CYC_CNT_DIVISOR;
- + sysclk <<= FE_US_CYC_CNT_SHIFT;
- +
- + fe_w32((fe_r32(FE_FE_GLO_CFG) &
- + ~(FE_US_CYC_CNT_MASK << FE_US_CYC_CNT_SHIFT)) |
- + sysclk,
- + FE_FE_GLO_CFG);
- + return 0;
- +}
- +
- +void fe_fwd_config(struct fe_priv *priv)
- +{
- + u32 fwd_cfg;
- +
- + fwd_cfg = fe_r32(FE_GDMA1_FWD_CFG);
- +
- + /* disable jumbo frame */
- + if (priv->flags & FE_FLAG_JUMBO_FRAME)
- + fwd_cfg &= ~FE_GDM1_JMB_EN;
- +
- + /* set unicast/multicast/broadcast frame to cpu */
- + fwd_cfg &= ~0xffff;
- +
- + fe_w32(fwd_cfg, FE_GDMA1_FWD_CFG);
- +}
- +
- +static void fe_rxcsum_config(bool enable)
- +{
- + if (enable)
- + fe_w32(fe_r32(FE_GDMA1_FWD_CFG) | (FE_GDM1_ICS_EN |
- + FE_GDM1_TCS_EN | FE_GDM1_UCS_EN),
- + FE_GDMA1_FWD_CFG);
- + else
- + fe_w32(fe_r32(FE_GDMA1_FWD_CFG) & ~(FE_GDM1_ICS_EN |
- + FE_GDM1_TCS_EN | FE_GDM1_UCS_EN),
- + FE_GDMA1_FWD_CFG);
- +}
- +
- +static void fe_txcsum_config(bool enable)
- +{
- + if (enable)
- + fe_w32(fe_r32(FE_CDMA_CSG_CFG) | (FE_ICS_GEN_EN |
- + FE_TCS_GEN_EN | FE_UCS_GEN_EN),
- + FE_CDMA_CSG_CFG);
- + else
- + fe_w32(fe_r32(FE_CDMA_CSG_CFG) & ~(FE_ICS_GEN_EN |
- + FE_TCS_GEN_EN | FE_UCS_GEN_EN),
- + FE_CDMA_CSG_CFG);
- +}
- +
- +void fe_csum_config(struct fe_priv *priv)
- +{
- + struct net_device *dev = priv_netdev(priv);
- +
- + fe_txcsum_config((dev->features & NETIF_F_IP_CSUM));
- + fe_rxcsum_config((dev->features & NETIF_F_RXCSUM));
- +}
- +
- +static int fe_hw_init(struct net_device *dev)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- + int i, err;
- +
- + err = devm_request_irq(priv->device, dev->irq, fe_handle_irq, 0,
- + dev_name(priv->device), dev);
- + if (err)
- + return err;
- +
- + if (priv->soc->set_mac)
- + priv->soc->set_mac(priv, dev->dev_addr);
- + else
- + fe_hw_set_macaddr(priv, dev->dev_addr);
- +
- + /* disable delay interrupt */
- + fe_reg_w32(0, FE_REG_DLY_INT_CFG);
- +
- + fe_int_disable(priv->soc->tx_int | priv->soc->rx_int);
- +
- + /* frame engine will push VLAN tag regarding to VIDX feild in Tx desc */
- + if (fe_reg_table[FE_REG_FE_DMA_VID_BASE])
- + for (i = 0; i < 16; i += 2)
- + fe_w32(((i + 1) << 16) + i,
- + fe_reg_table[FE_REG_FE_DMA_VID_BASE] +
- + (i * 2));
- +
- + if (priv->soc->fwd_config(priv))
- + netdev_err(dev, "unable to get clock\n");
- +
- + if (fe_reg_table[FE_REG_FE_RST_GL]) {
- + fe_reg_w32(1, FE_REG_FE_RST_GL);
- + fe_reg_w32(0, FE_REG_FE_RST_GL);
- + }
- +
- + return 0;
- +}
- +
- +static int fe_open(struct net_device *dev)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- + unsigned long flags;
- + u32 val;
- + int err;
- +
- + err = fe_init_dma(priv);
- + if (err) {
- + fe_free_dma(priv);
- + return err;
- + }
- +
- + spin_lock_irqsave(&priv->page_lock, flags);
- +
- + val = FE_TX_WB_DDONE | FE_RX_DMA_EN | FE_TX_DMA_EN;
- + if (priv->flags & FE_FLAG_RX_2B_OFFSET)
- + val |= FE_RX_2B_OFFSET;
- + val |= priv->soc->pdma_glo_cfg;
- + fe_reg_w32(val, FE_REG_PDMA_GLO_CFG);
- +
- + spin_unlock_irqrestore(&priv->page_lock, flags);
- +
- + if (priv->phy)
- + priv->phy->start(priv);
- +
- + if (priv->soc->has_carrier && priv->soc->has_carrier(priv))
- + netif_carrier_on(dev);
- +
- + napi_enable(&priv->rx_napi);
- + fe_int_enable(priv->soc->tx_int | priv->soc->rx_int);
- + netif_start_queue(dev);
- +
- + return 0;
- +}
- +
- +static int fe_stop(struct net_device *dev)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- + unsigned long flags;
- + int i;
- +
- + netif_tx_disable(dev);
- + fe_int_disable(priv->soc->tx_int | priv->soc->rx_int);
- + napi_disable(&priv->rx_napi);
- +
- + if (priv->phy)
- + priv->phy->stop(priv);
- +
- + spin_lock_irqsave(&priv->page_lock, flags);
- +
- + fe_reg_w32(fe_reg_r32(FE_REG_PDMA_GLO_CFG) &
- + ~(FE_TX_WB_DDONE | FE_RX_DMA_EN | FE_TX_DMA_EN),
- + FE_REG_PDMA_GLO_CFG);
- + spin_unlock_irqrestore(&priv->page_lock, flags);
- +
- + /* wait dma stop */
- + for (i = 0; i < 10; i++) {
- + if (fe_reg_r32(FE_REG_PDMA_GLO_CFG) &
- + (FE_TX_DMA_BUSY | FE_RX_DMA_BUSY)) {
- + msleep(20);
- + continue;
- + }
- + break;
- + }
- +
- + fe_free_dma(priv);
- +
- + return 0;
- +}
- +
- +static int __init fe_init(struct net_device *dev)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- + struct device_node *port;
- + const char *mac_addr;
- + int err;
- +
- + priv->soc->reset_fe();
- +
- + if (priv->soc->switch_init)
- + if (priv->soc->switch_init(priv)) {
- + netdev_err(dev, "failed to initialize switch core\n");
- + return -ENODEV;
- + }
- +
- + mac_addr = of_get_mac_address(priv->device->of_node);
- + if (mac_addr)
- + ether_addr_copy(dev->dev_addr, mac_addr);
- +
- + /* If the mac address is invalid, use random mac address */
- + if (!is_valid_ether_addr(dev->dev_addr)) {
- + random_ether_addr(dev->dev_addr);
- + dev_err(priv->device, "generated random MAC address %pM\n",
- + dev->dev_addr);
- + }
- +
- + err = fe_mdio_init(priv);
- + if (err)
- + return err;
- +
- + if (priv->soc->port_init)
- + for_each_child_of_node(priv->device->of_node, port)
- + if (of_device_is_compatible(port, "mediatek,eth-port") &&
- + of_device_is_available(port))
- + priv->soc->port_init(priv, port);
- +
- + if (priv->phy) {
- + err = priv->phy->connect(priv);
- + if (err)
- + goto err_phy_disconnect;
- + }
- +
- + err = fe_hw_init(dev);
- + if (!err)
- + return 0;
- +
- +err_phy_disconnect:
- + if (priv->phy)
- + priv->phy->disconnect(priv);
- + fe_mdio_cleanup(priv);
- +
- + return err;
- +}
- +
- +static void fe_uninit(struct net_device *dev)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- +
- + if (priv->phy)
- + priv->phy->disconnect(priv);
- + fe_mdio_cleanup(priv);
- +
- + fe_reg_w32(0, FE_REG_FE_INT_ENABLE);
- + free_irq(dev->irq, dev);
- +}
- +
- +static int fe_do_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- +
- + if (!priv->phy_dev)
- + return -ENODEV;
- +
- + switch (cmd) {
- + case SIOCGMIIPHY:
- + case SIOCGMIIREG:
- + case SIOCSMIIREG:
- + return phy_mii_ioctl(priv->phy_dev, ifr, cmd);
- + default:
- + break;
- + }
- +
- + return -EOPNOTSUPP;
- +}
- +
- +static int fe_change_mtu(struct net_device *dev, int new_mtu)
- +{
- + struct fe_priv *priv = netdev_priv(dev);
- + int frag_size, old_mtu;
- + u32 fwd_cfg;
- +
- + if (!(priv->flags & FE_FLAG_JUMBO_FRAME))
- + return eth_change_mtu(dev, new_mtu);
- +
- + frag_size = fe_max_frag_size(new_mtu);
- + if (new_mtu < 68 || frag_size > PAGE_SIZE)
- + return -EINVAL;
- +
- + old_mtu = dev->mtu;
- + dev->mtu = new_mtu;
- +
- + /* return early if the buffer sizes will not change */
- + if (old_mtu <= ETH_DATA_LEN && new_mtu <= ETH_DATA_LEN)
- + return 0;
- + if (old_mtu > ETH_DATA_LEN && new_mtu > ETH_DATA_LEN)
- + return 0;
- +
- + if (new_mtu <= ETH_DATA_LEN)
- + priv->rx_ring.frag_size = fe_max_frag_size(ETH_DATA_LEN);
- + else
- + priv->rx_ring.frag_size = PAGE_SIZE;
- + priv->rx_ring.rx_buf_size = fe_max_buf_size(priv->rx_ring.frag_size);
- +
- + if (!netif_running(dev))
- + return 0;
- +
- + fe_stop(dev);
- + fwd_cfg = fe_r32(FE_GDMA1_FWD_CFG);
- + if (new_mtu <= ETH_DATA_LEN) {
- + fwd_cfg &= ~FE_GDM1_JMB_EN;
- + } else {
- + fwd_cfg &= ~(FE_GDM1_JMB_LEN_MASK << FE_GDM1_JMB_LEN_SHIFT);
- + fwd_cfg |= (DIV_ROUND_UP(frag_size, 1024) <<
- + FE_GDM1_JMB_LEN_SHIFT) | FE_GDM1_JMB_EN;
- + }
- + fe_w32(fwd_cfg, FE_GDMA1_FWD_CFG);
- +
- + return fe_open(dev);
- +}
- +
- +static const struct net_device_ops fe_netdev_ops = {
- + .ndo_init = fe_init,
- + .ndo_uninit = fe_uninit,
- + .ndo_open = fe_open,
- + .ndo_stop = fe_stop,
- + .ndo_start_xmit = fe_start_xmit,
- + .ndo_set_mac_address = fe_set_mac_address,
- + .ndo_validate_addr = eth_validate_addr,
- + .ndo_do_ioctl = fe_do_ioctl,
- + .ndo_change_mtu = fe_change_mtu,
- + .ndo_tx_timeout = fe_tx_timeout,
- + .ndo_get_stats64 = fe_get_stats64,
- + .ndo_vlan_rx_add_vid = fe_vlan_rx_add_vid,
- + .ndo_vlan_rx_kill_vid = fe_vlan_rx_kill_vid,
- +#ifdef CONFIG_NET_POLL_CONTROLLER
- + .ndo_poll_controller = fe_poll_controller,
- +#endif
- +};
- +
- +static void fe_reset_pending(struct fe_priv *priv)
- +{
- + struct net_device *dev = priv->netdev;
- + int err;
- +
- + rtnl_lock();
- + fe_stop(dev);
- +
- + err = fe_open(dev);
- + if (err) {
- + netif_alert(priv, ifup, dev,
- + "Driver up/down cycle failed, closing device.\n");
- + dev_close(dev);
- + }
- + rtnl_unlock();
- +}
- +
- +static const struct fe_work_t fe_work[] = {
- + {FE_FLAG_RESET_PENDING, fe_reset_pending},
- +};
- +
- +static void fe_pending_work(struct work_struct *work)
- +{
- + struct fe_priv *priv = container_of(work, struct fe_priv, pending_work);
- + int i;
- + bool pending;
- +
- + for (i = 0; i < ARRAY_SIZE(fe_work); i++) {
- + pending = test_and_clear_bit(fe_work[i].bitnr,
- + priv->pending_flags);
- + if (pending)
- + fe_work[i].action(priv);
- + }
- +}
- +
- +static int fe_probe(struct platform_device *pdev)
- +{
- + struct resource *res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
- + const struct of_device_id *match;
- + struct fe_soc_data *soc;
- + struct net_device *netdev;
- + struct fe_priv *priv;
- + struct clk *sysclk;
- + int err, napi_weight;
- +
- + device_reset(&pdev->dev);
- +
- + match = of_match_device(of_fe_match, &pdev->dev);
- + soc = (struct fe_soc_data *)match->data;
- +
- + if (soc->reg_table)
- + fe_reg_table = soc->reg_table;
- + else
- + soc->reg_table = fe_reg_table;
- +
- + fe_base = devm_ioremap_resource(&pdev->dev, res);
- + if (!fe_base) {
- + err = -EADDRNOTAVAIL;
- + goto err_out;
- + }
- +
- + netdev = alloc_etherdev(sizeof(*priv));
- + if (!netdev) {
- + dev_err(&pdev->dev, "alloc_etherdev failed\n");
- + err = -ENOMEM;
- + goto err_iounmap;
- + }
- +
- + SET_NETDEV_DEV(netdev, &pdev->dev);
- + netdev->netdev_ops = &fe_netdev_ops;
- + netdev->base_addr = (unsigned long)fe_base;
- +
- + netdev->irq = platform_get_irq(pdev, 0);
- + if (netdev->irq < 0) {
- + dev_err(&pdev->dev, "no IRQ resource found\n");
- + err = -ENXIO;
- + goto err_free_dev;
- + }
- +
- + if (soc->init_data)
- + soc->init_data(soc, netdev);
- + netdev->vlan_features = netdev->hw_features & ~NETIF_F_HW_VLAN_CTAG_TX;
- + netdev->features |= netdev->hw_features;
- +
- + /* fake rx vlan filter func. to support tx vlan offload func */
- + if (fe_reg_table[FE_REG_FE_DMA_VID_BASE])
- + netdev->features |= NETIF_F_HW_VLAN_CTAG_FILTER;
- +
- + priv = netdev_priv(netdev);
- + spin_lock_init(&priv->page_lock);
- + if (fe_reg_table[FE_REG_FE_COUNTER_BASE]) {
- + priv->hw_stats = kzalloc(sizeof(*priv->hw_stats), GFP_KERNEL);
- + if (!priv->hw_stats) {
- + err = -ENOMEM;
- + goto err_free_dev;
- + }
- + spin_lock_init(&priv->hw_stats->stats_lock);
- + }
- +
- + sysclk = devm_clk_get(&pdev->dev, NULL);
- + if (!IS_ERR(sysclk)) {
- + priv->sysclk = clk_get_rate(sysclk);
- + } else if ((priv->flags & FE_FLAG_CALIBRATE_CLK)) {
- + dev_err(&pdev->dev, "this soc needs a clk for calibration\n");
- + err = -ENXIO;
- + goto err_free_dev;
- + }
- +
- + priv->switch_np = of_parse_phandle(pdev->dev.of_node, "mediatek,switch", 0);
- + if ((priv->flags & FE_FLAG_HAS_SWITCH) && !priv->switch_np) {
- + dev_err(&pdev->dev, "failed to read switch phandle\n");
- + err = -ENODEV;
- + goto err_free_dev;
- + }
- +
- + priv->netdev = netdev;
- + priv->device = &pdev->dev;
- + priv->soc = soc;
- + priv->msg_enable = netif_msg_init(fe_msg_level, FE_DEFAULT_MSG_ENABLE);
- + priv->rx_ring.frag_size = fe_max_frag_size(ETH_DATA_LEN);
- + priv->rx_ring.rx_buf_size = fe_max_buf_size(priv->rx_ring.frag_size);
- + priv->tx_ring.tx_ring_size = NUM_DMA_DESC;
- + priv->rx_ring.rx_ring_size = NUM_DMA_DESC;
- + INIT_WORK(&priv->pending_work, fe_pending_work);
- +
- + napi_weight = 32;
- + if (priv->flags & FE_FLAG_NAPI_WEIGHT) {
- + napi_weight *= 4;
- + priv->tx_ring.tx_ring_size *= 4;
- + priv->rx_ring.rx_ring_size *= 4;
- + }
- + netif_napi_add(netdev, &priv->rx_napi, fe_poll, napi_weight);
- + fe_set_ethtool_ops(netdev);
- +
- + err = register_netdev(netdev);
- + if (err) {
- + dev_err(&pdev->dev, "error bringing up device\n");
- + goto err_free_dev;
- + }
- +
- + platform_set_drvdata(pdev, netdev);
- +
- + netif_info(priv, probe, netdev, "mediatek frame engine at 0x%08lx, irq %d\n",
- + netdev->base_addr, netdev->irq);
- +
- + return 0;
- +
- +err_free_dev:
- + free_netdev(netdev);
- +err_iounmap:
- + devm_iounmap(&pdev->dev, fe_base);
- +err_out:
- + return err;
- +}
- +
- +static int fe_remove(struct platform_device *pdev)
- +{
- + struct net_device *dev = platform_get_drvdata(pdev);
- + struct fe_priv *priv = netdev_priv(dev);
- +
- + netif_napi_del(&priv->rx_napi);
- + kfree(priv->hw_stats);
- +
- + cancel_work_sync(&priv->pending_work);
- +
- + unregister_netdev(dev);
- + free_netdev(dev);
- + platform_set_drvdata(pdev, NULL);
- +
- + return 0;
- +}
- +
- +static struct platform_driver fe_driver = {
- + .probe = fe_probe,
- + .remove = fe_remove,
- + .driver = {
- + .name = "mtk_soc_eth",
- + .owner = THIS_MODULE,
- + .of_match_table = of_fe_match,
- + },
- +};
- +
- +module_platform_driver(fe_driver);
- +
- +MODULE_LICENSE("GPL");
- +MODULE_AUTHOR("John Crispin <blogic@openwrt.org>");
- +MODULE_DESCRIPTION("Ethernet driver for Ralink SoC");
- +MODULE_VERSION(MTK_FE_DRV_VERSION);
- --- /dev/null
- +++ b/drivers/net/ethernet/mediatek/mtk_eth_soc.h
- @@ -0,0 +1,522 @@
- +/* This program is free software; you can redistribute it and/or modify
- + * it under the terms of the GNU General Public License as published by
- + * the Free Software Foundation; version 2 of the License
- + *
- + * This program is distributed in the hope that it will be useful,
- + * but WITHOUT ANY WARRANTY; without even the implied warranty of
- + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- + * GNU General Public License for more details.
- + *
- + * Copyright (C) 2009-2015 John Crispin <blogic@openwrt.org>
- + * Copyright (C) 2009-2015 Felix Fietkau <nbd@nbd.name>
- + * Copyright (C) 2013-2015 Michael Lee <igvtee@gmail.com>
- + */
- +
- +#ifndef FE_ETH_H
- +#define FE_ETH_H
- +
- +#include <linux/mii.h>
- +#include <linux/interrupt.h>
- +#include <linux/netdevice.h>
- +#include <linux/dma-mapping.h>
- +#include <linux/phy.h>
- +#include <linux/ethtool.h>
- +#include <linux/version.h>
- +
- +enum fe_reg {
- + FE_REG_PDMA_GLO_CFG = 0,
- + FE_REG_PDMA_RST_CFG,
- + FE_REG_DLY_INT_CFG,
- + FE_REG_TX_BASE_PTR0,
- + FE_REG_TX_MAX_CNT0,
- + FE_REG_TX_CTX_IDX0,
- + FE_REG_TX_DTX_IDX0,
- + FE_REG_RX_BASE_PTR0,
- + FE_REG_RX_MAX_CNT0,
- + FE_REG_RX_CALC_IDX0,
- + FE_REG_RX_DRX_IDX0,
- + FE_REG_FE_INT_ENABLE,
- + FE_REG_FE_INT_STATUS,
- + FE_REG_FE_DMA_VID_BASE,
- + FE_REG_FE_COUNTER_BASE,
- + FE_REG_FE_RST_GL,
- + FE_REG_FE_INT_STATUS2,
- + FE_REG_COUNT
- +};
- +
- +enum fe_work_flag {
- + FE_FLAG_RESET_PENDING,
- + FE_FLAG_MAX
- +};
- +
- +#define MTK_FE_DRV_VERSION "0.1.2"
- +
- +/* power of 2 to let NEXT_TX_DESP_IDX work */
- +#define NUM_DMA_DESC BIT(7)
- +#define MAX_DMA_DESC 0xfff
- +
- +#define FE_DELAY_EN_INT 0x80
- +#define FE_DELAY_MAX_INT 0x04
- +#define FE_DELAY_MAX_TOUT 0x04
- +#define FE_DELAY_TIME 20
- +#define FE_DELAY_CHAN (((FE_DELAY_EN_INT | FE_DELAY_MAX_INT) << 8) | \
- + FE_DELAY_MAX_TOUT)
- +#define FE_DELAY_INIT ((FE_DELAY_CHAN << 16) | FE_DELAY_CHAN)
- +#define FE_PSE_FQFC_CFG_INIT 0x80504000
- +#define FE_PSE_FQFC_CFG_256Q 0xff908000
- +
- +/* interrupt bits */
- +#define FE_CNT_PPE_AF BIT(31)
- +#define FE_CNT_GDM_AF BIT(29)
- +#define FE_PSE_P2_FC BIT(26)
- +#define FE_PSE_BUF_DROP BIT(24)
- +#define FE_GDM_OTHER_DROP BIT(23)
- +#define FE_PSE_P1_FC BIT(22)
- +#define FE_PSE_P0_FC BIT(21)
- +#define FE_PSE_FQ_EMPTY BIT(20)
- +#define FE_GE1_STA_CHG BIT(18)
- +#define FE_TX_COHERENT BIT(17)
- +#define FE_RX_COHERENT BIT(16)
- +#define FE_TX_DONE_INT3 BIT(11)
- +#define FE_TX_DONE_INT2 BIT(10)
- +#define FE_TX_DONE_INT1 BIT(9)
- +#define FE_TX_DONE_INT0 BIT(8)
- +#define FE_RX_DONE_INT0 BIT(2)
- +#define FE_TX_DLY_INT BIT(1)
- +#define FE_RX_DLY_INT BIT(0)
- +
- +#define FE_RX_DONE_INT FE_RX_DONE_INT0
- +#define FE_TX_DONE_INT (FE_TX_DONE_INT0 | FE_TX_DONE_INT1 | \
- + FE_TX_DONE_INT2 | FE_TX_DONE_INT3)
- +
- +#define RT5350_RX_DLY_INT BIT(30)
- +#define RT5350_TX_DLY_INT BIT(28)
- +#define RT5350_RX_DONE_INT1 BIT(17)
- +#define RT5350_RX_DONE_INT0 BIT(16)
- +#define RT5350_TX_DONE_INT3 BIT(3)
- +#define RT5350_TX_DONE_INT2 BIT(2)
- +#define RT5350_TX_DONE_INT1 BIT(1)
- +#define RT5350_TX_DONE_INT0 BIT(0)
- +
- +#define RT5350_RX_DONE_INT (RT5350_RX_DONE_INT0 | RT5350_RX_DONE_INT1)
- +#define RT5350_TX_DONE_INT (RT5350_TX_DONE_INT0 | RT5350_TX_DONE_INT1 | \
- + RT5350_TX_DONE_INT2 | RT5350_TX_DONE_INT3)
- +
- +/* registers */
- +#define FE_FE_OFFSET 0x0000
- +#define FE_GDMA_OFFSET 0x0020
- +#define FE_PSE_OFFSET 0x0040
- +#define FE_GDMA2_OFFSET 0x0060
- +#define FE_CDMA_OFFSET 0x0080
- +#define FE_DMA_VID0 0x00a8
- +#define FE_PDMA_OFFSET 0x0100
- +#define FE_PPE_OFFSET 0x0200
- +#define FE_CMTABLE_OFFSET 0x0400
- +#define FE_POLICYTABLE_OFFSET 0x1000
- +
- +#define RT5350_PDMA_OFFSET 0x0800
- +#define RT5350_SDM_OFFSET 0x0c00
- +
- +#define FE_MDIO_ACCESS (FE_FE_OFFSET + 0x00)
- +#define FE_MDIO_CFG (FE_FE_OFFSET + 0x04)
- +#define FE_FE_GLO_CFG (FE_FE_OFFSET + 0x08)
- +#define FE_FE_RST_GL (FE_FE_OFFSET + 0x0C)
- +#define FE_FE_INT_STATUS (FE_FE_OFFSET + 0x10)
- +#define FE_FE_INT_ENABLE (FE_FE_OFFSET + 0x14)
- +#define FE_MDIO_CFG2 (FE_FE_OFFSET + 0x18)
- +#define FE_FOC_TS_T (FE_FE_OFFSET + 0x1C)
- +
- +#define FE_GDMA1_FWD_CFG (FE_GDMA_OFFSET + 0x00)
- +#define FE_GDMA1_SCH_CFG (FE_GDMA_OFFSET + 0x04)
- +#define FE_GDMA1_SHPR_CFG (FE_GDMA_OFFSET + 0x08)
- +#define FE_GDMA1_MAC_ADRL (FE_GDMA_OFFSET + 0x0C)
- +#define FE_GDMA1_MAC_ADRH (FE_GDMA_OFFSET + 0x10)
- +
- +#define FE_GDMA2_FWD_CFG (FE_GDMA2_OFFSET + 0x00)
- +#define FE_GDMA2_SCH_CFG (FE_GDMA2_OFFSET + 0x04)
- +#define FE_GDMA2_SHPR_CFG (FE_GDMA2_OFFSET + 0x08)
- +#define FE_GDMA2_MAC_ADRL (FE_GDMA2_OFFSET + 0x0C)
- +#define FE_GDMA2_MAC_ADRH (FE_GDMA2_OFFSET + 0x10)
- +
- +#define FE_PSE_FQ_CFG (FE_PSE_OFFSET + 0x00)
- +#define FE_CDMA_FC_CFG (FE_PSE_OFFSET + 0x04)
- +#define FE_GDMA1_FC_CFG (FE_PSE_OFFSET + 0x08)
- +#define FE_GDMA2_FC_CFG (FE_PSE_OFFSET + 0x0C)
- +
- +#define FE_CDMA_CSG_CFG (FE_CDMA_OFFSET + 0x00)
- +#define FE_CDMA_SCH_CFG (FE_CDMA_OFFSET + 0x04)
- +
- +#ifdef CONFIG_SOC_MT7621
- +#define MT7620A_GDMA_OFFSET 0x0500
- +#else
- +#define MT7620A_GDMA_OFFSET 0x0600
- +#endif
- +#define MT7620A_GDMA1_FWD_CFG (MT7620A_GDMA_OFFSET + 0x00)
- +#define MT7620A_FE_GDMA1_SCH_CFG (MT7620A_GDMA_OFFSET + 0x04)
- +#define MT7620A_FE_GDMA1_SHPR_CFG (MT7620A_GDMA_OFFSET + 0x08)
- +#define MT7620A_FE_GDMA1_MAC_ADRL (MT7620A_GDMA_OFFSET + 0x0C)
- +#define MT7620A_FE_GDMA1_MAC_ADRH (MT7620A_GDMA_OFFSET + 0x10)
- +
- +#define RT5350_TX_BASE_PTR0 (RT5350_PDMA_OFFSET + 0x00)
- +#define RT5350_TX_MAX_CNT0 (RT5350_PDMA_OFFSET + 0x04)
- +#define RT5350_TX_CTX_IDX0 (RT5350_PDMA_OFFSET + 0x08)
- +#define RT5350_TX_DTX_IDX0 (RT5350_PDMA_OFFSET + 0x0C)
- +#define RT5350_TX_BASE_PTR1 (RT5350_PDMA_OFFSET + 0x10)
- +#define RT5350_TX_MAX_CNT1 (RT5350_PDMA_OFFSET + 0x14)
- +#define RT5350_TX_CTX_IDX1 (RT5350_PDMA_OFFSET + 0x18)
- +#define RT5350_TX_DTX_IDX1 (RT5350_PDMA_OFFSET + 0x1C)
- +#define RT5350_TX_BASE_PTR2 (RT5350_PDMA_OFFSET + 0x20)
- +#define RT5350_TX_MAX_CNT2 (RT5350_PDMA_OFFSET + 0x24)
- +#define RT5350_TX_CTX_IDX2 (RT5350_PDMA_OFFSET + 0x28)
- +#define RT5350_TX_DTX_IDX2 (RT5350_PDMA_OFFSET + 0x2C)
- +#define RT5350_TX_BASE_PTR3 (RT5350_PDMA_OFFSET + 0x30)
- +#define RT5350_TX_MAX_CNT3 (RT5350_PDMA_OFFSET + 0x34)
- +#define RT5350_TX_CTX_IDX3 (RT5350_PDMA_OFFSET + 0x38)
- +#define RT5350_TX_DTX_IDX3 (RT5350_PDMA_OFFSET + 0x3C)
- +#define RT5350_RX_BASE_PTR0 (RT5350_PDMA_OFFSET + 0x100)
- +#define RT5350_RX_MAX_CNT0 (RT5350_PDMA_OFFSET + 0x104)
- +#define RT5350_RX_CALC_IDX0 (RT5350_PDMA_OFFSET + 0x108)
- +#define RT5350_RX_DRX_IDX0 (RT5350_PDMA_OFFSET + 0x10C)
- +#define RT5350_RX_BASE_PTR1 (RT5350_PDMA_OFFSET + 0x110)
- +#define RT5350_RX_MAX_CNT1 (RT5350_PDMA_OFFSET + 0x114)
- +#define RT5350_RX_CALC_IDX1 (RT5350_PDMA_OFFSET + 0x118)
- +#define RT5350_RX_DRX_IDX1 (RT5350_PDMA_OFFSET + 0x11C)
- +#define RT5350_PDMA_GLO_CFG (RT5350_PDMA_OFFSET + 0x204)
- +#define RT5350_PDMA_RST_CFG (RT5350_PDMA_OFFSET + 0x208)
- +#define RT5350_DLY_INT_CFG (RT5350_PDMA_OFFSET + 0x20c)
- +#define RT5350_FE_INT_STATUS (RT5350_PDMA_OFFSET + 0x220)
- +#define RT5350_FE_INT_ENABLE (RT5350_PDMA_OFFSET + 0x228)
- +#define RT5350_PDMA_SCH_CFG (RT5350_PDMA_OFFSET + 0x280)
- +
- +#define FE_PDMA_GLO_CFG (FE_PDMA_OFFSET + 0x00)
- +#define FE_PDMA_RST_CFG (FE_PDMA_OFFSET + 0x04)
- +#define FE_PDMA_SCH_CFG (FE_PDMA_OFFSET + 0x08)
- +#define FE_DLY_INT_CFG (FE_PDMA_OFFSET + 0x0C)
- +#define FE_TX_BASE_PTR0 (FE_PDMA_OFFSET + 0x10)
- +#define FE_TX_MAX_CNT0 (FE_PDMA_OFFSET + 0x14)
- +#define FE_TX_CTX_IDX0 (FE_PDMA_OFFSET + 0x18)
- +#define FE_TX_DTX_IDX0 (FE_PDMA_OFFSET + 0x1C)
- +#define FE_TX_BASE_PTR1 (FE_PDMA_OFFSET + 0x20)
- +#define FE_TX_MAX_CNT1 (FE_PDMA_OFFSET + 0x24)
- +#define FE_TX_CTX_IDX1 (FE_PDMA_OFFSET + 0x28)
- +#define FE_TX_DTX_IDX1 (FE_PDMA_OFFSET + 0x2C)
- +#define FE_RX_BASE_PTR0 (FE_PDMA_OFFSET + 0x30)
- +#define FE_RX_MAX_CNT0 (FE_PDMA_OFFSET + 0x34)
- +#define FE_RX_CALC_IDX0 (FE_PDMA_OFFSET + 0x38)
- +#define FE_RX_DRX_IDX0 (FE_PDMA_OFFSET + 0x3C)
- +#define FE_TX_BASE_PTR2 (FE_PDMA_OFFSET + 0x40)
- +#define FE_TX_MAX_CNT2 (FE_PDMA_OFFSET + 0x44)
- +#define FE_TX_CTX_IDX2 (FE_PDMA_OFFSET + 0x48)
- +#define FE_TX_DTX_IDX2 (FE_PDMA_OFFSET + 0x4C)
- +#define FE_TX_BASE_PTR3 (FE_PDMA_OFFSET + 0x50)
- +#define FE_TX_MAX_CNT3 (FE_PDMA_OFFSET + 0x54)
- +#define FE_TX_CTX_IDX3 (FE_PDMA_OFFSET + 0x58)
- +#define FE_TX_DTX_IDX3 (FE_PDMA_OFFSET + 0x5C)
- +#define FE_RX_BASE_PTR1 (FE_PDMA_OFFSET + 0x60)
- +#define FE_RX_MAX_CNT1 (FE_PDMA_OFFSET + 0x64)
- +#define FE_RX_CALC_IDX1 (FE_PDMA_OFFSET + 0x68)
- +#define FE_RX_DRX_IDX1 (FE_PDMA_OFFSET + 0x6C)
- +
- +/* Switch DMA configuration */
- +#define RT5350_SDM_CFG (RT5350_SDM_OFFSET + 0x00)
- +#define RT5350_SDM_RRING (RT5350_SDM_OFFSET + 0x04)
- +#define RT5350_SDM_TRING (RT5350_SDM_OFFSET + 0x08)
- +#define RT5350_SDM_MAC_ADRL (RT5350_SDM_OFFSET + 0x0C)
- +#define RT5350_SDM_MAC_ADRH (RT5350_SDM_OFFSET + 0x10)
- +#define RT5350_SDM_TPCNT (RT5350_SDM_OFFSET + 0x100)
- +#define RT5350_SDM_TBCNT (RT5350_SDM_OFFSET + 0x104)
- +#define RT5350_SDM_RPCNT (RT5350_SDM_OFFSET + 0x108)
- +#define RT5350_SDM_RBCNT (RT5350_SDM_OFFSET + 0x10C)
- +#define RT5350_SDM_CS_ERR (RT5350_SDM_OFFSET + 0x110)
- +
- +#define RT5350_SDM_ICS_EN BIT(16)
- +#define RT5350_SDM_TCS_EN BIT(17)
- +#define RT5350_SDM_UCS_EN BIT(18)
- +
- +/* MDIO_CFG register bits */
- +#define FE_MDIO_CFG_AUTO_POLL_EN BIT(29)
- +#define FE_MDIO_CFG_GP1_BP_EN BIT(16)
- +#define FE_MDIO_CFG_GP1_FRC_EN BIT(15)
- +#define FE_MDIO_CFG_GP1_SPEED_10 (0 << 13)
- +#define FE_MDIO_CFG_GP1_SPEED_100 (1 << 13)
- +#define FE_MDIO_CFG_GP1_SPEED_1000 (2 << 13)
- +#define FE_MDIO_CFG_GP1_DUPLEX BIT(12)
- +#define FE_MDIO_CFG_GP1_FC_TX BIT(11)
- +#define FE_MDIO_CFG_GP1_FC_RX BIT(10)
- +#define FE_MDIO_CFG_GP1_LNK_DWN BIT(9)
- +#define FE_MDIO_CFG_GP1_AN_FAIL BIT(8)
- +#define FE_MDIO_CFG_MDC_CLK_DIV_1 (0 << 6)
- +#define FE_MDIO_CFG_MDC_CLK_DIV_2 (1 << 6)
- +#define FE_MDIO_CFG_MDC_CLK_DIV_4 (2 << 6)
- +#define FE_MDIO_CFG_MDC_CLK_DIV_8 (3 << 6)
- +#define FE_MDIO_CFG_TURBO_MII_FREQ BIT(5)
- +#define FE_MDIO_CFG_TURBO_MII_MODE BIT(4)
- +#define FE_MDIO_CFG_RX_CLK_SKEW_0 (0 << 2)
- +#define FE_MDIO_CFG_RX_CLK_SKEW_200 (1 << 2)
- +#define FE_MDIO_CFG_RX_CLK_SKEW_400 (2 << 2)
- +#define FE_MDIO_CFG_RX_CLK_SKEW_INV (3 << 2)
- +#define FE_MDIO_CFG_TX_CLK_SKEW_0 0
- +#define FE_MDIO_CFG_TX_CLK_SKEW_200 1
- +#define FE_MDIO_CFG_TX_CLK_SKEW_400 2
- +#define FE_MDIO_CFG_TX_CLK_SKEW_INV 3
- +
- +/* uni-cast port */
- +#define FE_GDM1_JMB_LEN_MASK 0xf
- +#define FE_GDM1_JMB_LEN_SHIFT 28
- +#define FE_GDM1_ICS_EN BIT(22)
- +#define FE_GDM1_TCS_EN BIT(21)
- +#define FE_GDM1_UCS_EN BIT(20)
- +#define FE_GDM1_JMB_EN BIT(19)
- +#define FE_GDM1_STRPCRC BIT(16)
- +#define FE_GDM1_UFRC_P_CPU (0 << 12)
- +#define FE_GDM1_UFRC_P_GDMA1 (1 << 12)
- +#define FE_GDM1_UFRC_P_PPE (6 << 12)
- +
- +/* checksums */
- +#define FE_ICS_GEN_EN BIT(2)
- +#define FE_UCS_GEN_EN BIT(1)
- +#define FE_TCS_GEN_EN BIT(0)
- +
- +/* dma ring */
- +#define FE_PST_DRX_IDX0 BIT(16)
- +#define FE_PST_DTX_IDX3 BIT(3)
- +#define FE_PST_DTX_IDX2 BIT(2)
- +#define FE_PST_DTX_IDX1 BIT(1)
- +#define FE_PST_DTX_IDX0 BIT(0)
- +
- +#define FE_RX_2B_OFFSET BIT(31)
- +#define FE_TX_WB_DDONE BIT(6)
- +#define FE_RX_DMA_BUSY BIT(3)
- +#define FE_TX_DMA_BUSY BIT(1)
- +#define FE_RX_DMA_EN BIT(2)
- +#define FE_TX_DMA_EN BIT(0)
- +
- +#define FE_PDMA_SIZE_4DWORDS (0 << 4)
- +#define FE_PDMA_SIZE_8DWORDS (1 << 4)
- +#define FE_PDMA_SIZE_16DWORDS (2 << 4)
- +
- +#define FE_US_CYC_CNT_MASK 0xff
- +#define FE_US_CYC_CNT_SHIFT 0x8
- +#define FE_US_CYC_CNT_DIVISOR 1000000
- +
- +/* rxd2 */
- +#define RX_DMA_DONE BIT(31)
- +#define RX_DMA_LSO BIT(30)
- +#define RX_DMA_PLEN0(_x) (((_x) & 0x3fff) << 16)
- +#define RX_DMA_GET_PLEN0(_x) (((_x) >> 16) & 0x3fff)
- +#define RX_DMA_TAG BIT(15)
- +/* rxd3 */
- +#define RX_DMA_TPID(_x) (((_x) >> 16) & 0xffff)
- +#define RX_DMA_VID(_x) ((_x) & 0xffff)
- +/* rxd4 */
- +#define RX_DMA_L4VALID BIT(30)
- +
- +struct fe_rx_dma {
- + unsigned int rxd1;
- + unsigned int rxd2;
- + unsigned int rxd3;
- + unsigned int rxd4;
- +} __packed __aligned(4);
- +
- +#define TX_DMA_BUF_LEN 0x3fff
- +#define TX_DMA_PLEN0_MASK (TX_DMA_BUF_LEN << 16)
- +#define TX_DMA_PLEN0(_x) (((_x) & TX_DMA_BUF_LEN) << 16)
- +#define TX_DMA_PLEN1(_x) ((_x) & TX_DMA_BUF_LEN)
- +#define TX_DMA_GET_PLEN0(_x) (((_x) >> 16) & TX_DMA_BUF_LEN)
- +#define TX_DMA_GET_PLEN1(_x) ((_x) & TX_DMA_BUF_LEN)
- +#define TX_DMA_LS1 BIT(14)
- +#define TX_DMA_LS0 BIT(30)
- +#define TX_DMA_DONE BIT(31)
- +
- +#define TX_DMA_INS_VLAN_MT7621 BIT(16)
- +#define TX_DMA_INS_VLAN BIT(7)
- +#define TX_DMA_INS_PPPOE BIT(12)
- +#define TX_DMA_QN(_x) ((_x) << 16)
- +#define TX_DMA_PN(_x) ((_x) << 24)
- +#define TX_DMA_QN_MASK TX_DMA_QN(0x7)
- +#define TX_DMA_PN_MASK TX_DMA_PN(0x7)
- +#define TX_DMA_UDF BIT(20)
- +#define TX_DMA_CHKSUM (0x7 << 29)
- +#define TX_DMA_TSO BIT(28)
- +
- +/* frame engine counters */
- +#define FE_PPE_AC_BCNT0 (FE_CMTABLE_OFFSET + 0x00)
- +#define FE_GDMA1_TX_GBCNT (FE_CMTABLE_OFFSET + 0x300)
- +#define FE_GDMA2_TX_GBCNT (FE_GDMA1_TX_GBCNT + 0x40)
- +
- +/* phy device flags */
- +#define FE_PHY_FLAG_PORT BIT(0)
- +#define FE_PHY_FLAG_ATTACH BIT(1)
- +
- +struct fe_tx_dma {
- + unsigned int txd1;
- + unsigned int txd2;
- + unsigned int txd3;
- + unsigned int txd4;
- +} __packed __aligned(4);
- +
- +struct fe_priv;
- +
- +struct fe_phy {
- + /* make sure that phy operations are atomic */
- + spinlock_t lock;
- +
- + struct phy_device *phy[8];
- + struct device_node *phy_node[8];
- + const __be32 *phy_fixed[8];
- + int duplex[8];
- + int speed[8];
- + int tx_fc[8];
- + int rx_fc[8];
- + int (*connect)(struct fe_priv *priv);
- + void (*disconnect)(struct fe_priv *priv);
- + void (*start)(struct fe_priv *priv);
- + void (*stop)(struct fe_priv *priv);
- +};
- +
- +struct fe_soc_data {
- + const u16 *reg_table;
- +
- + void (*init_data)(struct fe_soc_data *data, struct net_device *netdev);
- + void (*reset_fe)(void);
- + void (*set_mac)(struct fe_priv *priv, unsigned char *mac);
- + int (*fwd_config)(struct fe_priv *priv);
- + void (*tx_dma)(struct fe_tx_dma *txd);
- + int (*switch_init)(struct fe_priv *priv);
- + void (*port_init)(struct fe_priv *priv, struct device_node *port);
- + int (*has_carrier)(struct fe_priv *priv);
- + int (*mdio_init)(struct fe_priv *priv);
- + void (*mdio_cleanup)(struct fe_priv *priv);
- + int (*mdio_write)(struct mii_bus *bus, int phy_addr, int phy_reg,
- + u16 val);
- + int (*mdio_read)(struct mii_bus *bus, int phy_addr, int phy_reg);
- + void (*mdio_adjust_link)(struct fe_priv *priv, int port);
- +
- + void *swpriv;
- + u32 pdma_glo_cfg;
- + u32 rx_int;
- + u32 tx_int;
- + u32 status_int;
- + u32 checksum_bit;
- +};
- +
- +#define FE_FLAG_PADDING_64B BIT(0)
- +#define FE_FLAG_PADDING_BUG BIT(1)
- +#define FE_FLAG_JUMBO_FRAME BIT(2)
- +#define FE_FLAG_RX_2B_OFFSET BIT(3)
- +#define FE_FLAG_RX_SG_DMA BIT(4)
- +#define FE_FLAG_RX_VLAN_CTAG BIT(5)
- +#define FE_FLAG_NAPI_WEIGHT BIT(6)
- +#define FE_FLAG_CALIBRATE_CLK BIT(7)
- +#define FE_FLAG_HAS_SWITCH BIT(8)
- +
- +#define FE_STAT_REG_DECLARE \
- + _FE(tx_bytes) \
- + _FE(tx_packets) \
- + _FE(tx_skip) \
- + _FE(tx_collisions) \
- + _FE(rx_bytes) \
- + _FE(rx_packets) \
- + _FE(rx_overflow) \
- + _FE(rx_fcs_errors) \
- + _FE(rx_short_errors) \
- + _FE(rx_long_errors) \
- + _FE(rx_checksum_errors) \
- + _FE(rx_flow_control_packets)
- +
- +struct fe_hw_stats {
- + /* make sure that stats operations are atomic */
- + spinlock_t stats_lock;
- +
- + struct u64_stats_sync syncp;
- +#define _FE(x) u64 x;
- + FE_STAT_REG_DECLARE
- +#undef _FE
- +};
- +
- +enum fe_tx_flags {
- + FE_TX_FLAGS_SINGLE0 = 0x01,
- + FE_TX_FLAGS_PAGE0 = 0x02,
- + FE_TX_FLAGS_PAGE1 = 0x04,
- +};
- +
- +struct fe_tx_buf {
- + struct sk_buff *skb;
- + u32 flags;
- + DEFINE_DMA_UNMAP_ADDR(dma_addr0);
- + DEFINE_DMA_UNMAP_LEN(dma_len0);
- + DEFINE_DMA_UNMAP_ADDR(dma_addr1);
- + DEFINE_DMA_UNMAP_LEN(dma_len1);
- +};
- +
- +struct fe_tx_ring {
- + struct fe_tx_dma *tx_dma;
- + struct fe_tx_buf *tx_buf;
- + dma_addr_t tx_phys;
- + u16 tx_ring_size;
- + u16 tx_free_idx;
- + u16 tx_next_idx;
- + u16 tx_thresh;
- +};
- +
- +struct fe_rx_ring {
- + struct fe_rx_dma *rx_dma;
- + u8 **rx_data;
- + dma_addr_t rx_phys;
- + u16 rx_ring_size;
- + u16 frag_size;
- + u16 rx_buf_size;
- + u16 rx_calc_idx;
- +};
- +
- +struct fe_priv {
- + /* make sure that register operations are atomic */
- + spinlock_t page_lock;
- +
- + struct fe_soc_data *soc;
- + struct net_device *netdev;
- + struct device_node *switch_np;
- + u32 msg_enable;
- + u32 flags;
- +
- + struct device *device;
- + unsigned long sysclk;
- +
- + struct fe_rx_ring rx_ring;
- + struct napi_struct rx_napi;
- +
- + struct fe_tx_ring tx_ring;
- +
- + struct fe_phy *phy;
- + struct mii_bus *mii_bus;
- + struct phy_device *phy_dev;
- + u32 phy_flags;
- +
- + int link[8];
- +
- + struct fe_hw_stats *hw_stats;
- + unsigned long vlan_map;
- + struct work_struct pending_work;
- + DECLARE_BITMAP(pending_flags, FE_FLAG_MAX);
- +};
- +
- +extern const struct of_device_id of_fe_match[];
- +
- +void fe_w32(u32 val, unsigned reg);
- +u32 fe_r32(unsigned reg);
- +
- +int fe_set_clock_cycle(struct fe_priv *priv);
- +void fe_csum_config(struct fe_priv *priv);
- +void fe_stats_update(struct fe_priv *priv);
- +void fe_fwd_config(struct fe_priv *priv);
- +void fe_reg_w32(u32 val, enum fe_reg reg);
- +u32 fe_reg_r32(enum fe_reg reg);
- +
- +void fe_reset(u32 reset_bits);
- +
- +static inline void *priv_netdev(struct fe_priv *priv)
- +{
- + return (char *)priv - ALIGN(sizeof(struct net_device), NETDEV_ALIGN);
- +}
- +
- +#endif /* FE_ETH_H */
|