code
stringlengths
5
1.03M
repo_name
stringlengths
5
90
path
stringlengths
4
158
license
stringclasses
15 values
size
int64
5
1.03M
n_ast_errors
int64
0
53.9k
ast_max_depth
int64
2
4.17k
n_whitespaces
int64
0
365k
n_ast_nodes
int64
3
317k
n_ast_terminals
int64
1
171k
n_ast_nonterminals
int64
1
146k
loc
int64
-1
37.3k
cycloplexity
int64
-1
1.31k
module Tiling.A295229Spec (main, spec) where import Test.Hspec import Tiling.A295229 (a295229) main :: IO () main = hspec spec spec :: Spec spec = describe "A295229" $ it "correctly computes the first 10 elements" $ take 10 (map a295229 [1..]) `shouldBe` expectedValue where expectedValue = [1, 6, 84, 8548, 4203520, 8590557312, 70368815480832, 2305843028004192256, 302231454912728264605696, 158456325028538104598816096256]
peterokagey/haskellOEIS
test/Tiling/A295229Spec.hs
apache-2.0
438
0
10
68
130
75
55
10
1
module Permutations.A330858 (a330858, a330858T) where import Permutations.A068424 (a068424T) import Data.MemoCombinators (memo2, integral) a330858T :: Integer -> Integer -> Integer a330858T = memo2 integral integral a330858T' where a330858T' n k | n == 0 = 1 | n <= k = product [1..n] | otherwise = n * a330858T (n-1) k - a068424T (n-1) k * a330858T (n-k-1) k a330858_row :: Integer -> [Integer] a330858_row n = map (a330858T n) [1..n] a330858_list :: [Integer] a330858_list = concatMap a330858_row [1..] a330858 :: Int -> Integer a330858 n = a330858_list !! (n - 1)
peterokagey/haskellOEIS
src/Permutations/A330858.hs
apache-2.0
591
0
13
115
251
132
119
15
1
-- logging feature implemented with do syntax import Control.Monad.Writer -- be careful: the code does not compile correctly logNumber :: Int -> Writer [String] Int logNumber x = Writer (x, ["Got number: " ++ show x]) multWithLog :: Writer [String] Int multWithLog = do a <- logNumber 3 b <- logNumber 5 return (a*b)
Oscarzhao/haskell
learnyouahaskell/monads/log_with_do.hs
apache-2.0
323
0
9
60
104
53
51
8
1
{-# OPTIONS_GHC -fno-warn-name-shadowing #-} {----------------------------------------------------------------- This files provides functions to convert parsed JSON responses of type Jvalue to other Haskell Types that we use. All functions in this module are pure. So, all error reporting is done using (Either MercadoException a). No synchronous (imprecise) exception should ever be thrown from within this module. This is a tighter match than a naïve Aeson implementation. I have found a few bugs in their API through this tighter coupling: - Order of parameters is not important - For safety, these lists will be sorted (even if they already are): - asks are sorted by increasing price - bids are sorted by decreasing price I don't know what will happen to the trading algorithms if they are not! So, we must ensure that. - We also ensure that prices and volumes are non-negative numbers ------------------------------------------------------------------} module Mercado.Internals ( module Mercado.Internals , parse ) where import Text.Read (readMaybe) import Data.List import Data.Word import Data.Function import qualified Data.HashMap.Strict as Map import Razao.Util import JsonParser import Mercado.Types data MercadoAPIType = GetInfo | GetOrder | OrderList | CancelOrder | WithdrawCoin | PlaceSellOrder | PlaceBuyOrder | MercadoListBook instance Show MercadoAPIType where -- v3 show PlaceSellOrder = "place_sell_order" show PlaceBuyOrder = "place_buy_order" show MercadoListBook = "list_orderbook" show CancelOrder = "cancel_order" show OrderList = "list_orders" show GetInfo = "get_account_info" show GetOrder = "get_order" show WithdrawCoin = "withdraw_coin" {- little reminder: data Jvalue = Jobject (Map.HashMap String Jvalue) | Jarray [Jvalue] | Jstring String -- terminals | Jnumber Double | Jbool Bool | Jnull deriving Show parse :: String -> Jvalue -} ------------------------------------------------ getMercadoResponseTime :: Jvalue -> Maybe Timestamp getMercadoResponseTime (Jobject dict) = Map.lookup "server_unix_timestamp" dict >>= getString >>= readMaybe >>= maybePositive getMercadoResponseTime _ = Nothing ------------------------------------------------ type MercadoDataParser a = Timestamp -> Jvalue -> Either MercadoException a convertMercadoResponse :: MercadoDataParser a -> Jvalue -> (Either MercadoException a) convertMercadoResponse parser resp@(Jobject dict) = let mRespStatus = Map.lookup "status_code" dict mResponseData = Map.lookup "response_data" dict mTime = getMercadoResponseTime resp in case (mRespStatus, mTime, mResponseData) of (Just (Jnumber 100), Just t , Just jResponseData) -> parser t jResponseData (Just (Jnumber 100), Just _ , Nothing ) -> Left (BadResponseStructure $ "Missing response data: " ++ show resp) (Just (Jnumber 100), Nothing, _ ) -> Left (BadResponseStructure $ "Bad or missing server_unix_timestamp: " ++ show resp) (Nothing , _ , _ ) -> Left (BadResponseStructure $ "Missing status code: " ++ show resp) (Just (Jnumber 200), _ , _ ) -> Left (OtherError $ show resp) (Just (Jnumber 201), _ , _ ) -> Left (Invalid_TAPI_ID $ show resp) (Just (Jnumber 202), _ , _ ) -> Left (Invalid_TAPI_MAC $ show resp) (Just (Jnumber 203), _ , _ ) -> Left (Invalid_Tonce $ show resp) (Just (Jnumber 204), _ , _ ) -> Left (Invalid_TAPI_Method $ show resp) (Just (Jnumber 205), _ , _ ) -> Left (InvalidCoinPair $ show resp) (Just (Jnumber 206), _ , _ ) -> Left (InvalidParameter $ show resp) (Just (Jnumber 207), _ , _ ) -> Left (InsufficientFunds $ show resp) (Just (Jnumber 208), _ , _ ) -> Left (InvalidOrder $ show resp) (Just (Jnumber 209), _ , _ ) -> Left (UntrustedWalletAddress $ show resp) (Just (Jnumber 210), _ , _ ) -> Left (BRLWithdrawalFailure $ show resp) (Just (Jnumber 211), _ , _ ) -> Left (InvalidForReadOnlyKey $ show resp) (Just (Jnumber 212), _ , _ ) -> Left (OrderAlreadyDone $ show resp) (Just (Jnumber 213), _ , _ ) -> Left (UntrustedBankAccount $ show resp) (Just (Jnumber 214), _ , _ ) -> Left (OtherError $ show resp) (Just (Jnumber 215), _ , _ ) -> Left (NotEnoughBitcoins $ show resp) (Just (Jnumber 216), _ , _ ) -> Left (NotEnoughLitecoins $ show resp) (Just (Jnumber 217), _ , _ ) -> Left (BRLWithdrawalLimitReached $ show resp) (Just (Jnumber 218), _ , _ ) -> Left (BTCWithdrawalLimitReached $ show resp) (Just (Jnumber 219), _ , _ ) -> Left (OtherError $ show resp) (Just (Jnumber 220), _ , _ ) -> Left (BTCWithdrawalTooSmall $ show resp) (Just (Jnumber 221), _ , _ ) -> Left (LTCWithdrawalTooSmall $ show resp) (Just (Jnumber 222), _ , _ ) -> Left (BTCVolumeTooSmall $ show resp) (Just (Jnumber 223), _ , _ ) -> Left (LTCVolumeTooSmall $ show resp) (Just (Jnumber 224), _ , _ ) -> Left (BelowMiminumPrice $ show resp) (Just (Jnumber 225), _ , _ ) -> Left (OtherError $ show resp) (Just (Jnumber 226), _ , _ ) -> Left (OtherError $ show resp) (Just (Jnumber 227), _ , _ ) -> Left (TooMuchPrecision $ show resp) (Just (Jnumber 429), _ , _ ) -> Left (MaxRequestRateExceeded $ show resp) (Just (Jnumber 500), _ , _ ) -> Left (InternalServerError $ show resp) (Just (Jnumber _ ), _ , _ ) -> Left (OtherError $ "Unknown status code: " ++ show resp) (Just _ , _ , _ ) -> Left (BadResponseStructure $ "Bad status code: " ++ show resp) -- Finally, we have an error if response is not a dictionary convertMercadoResponse _parser resp = Left (BadResponseStructure $ "Response: " ++ show resp) badFormatWrapper :: (Timestamp -> Jvalue -> Maybe a) -> Timestamp -> Jvalue -> Either MercadoException a badFormatWrapper parser time jval = case parser time jval of Nothing -> Left $ BadResponseStructure (show jval) Just x -> Right $ x ------------------------------------------------ -- functions to convert orderbooks convertMercadoBook :: (Ord p, Fractional p, Fractional v) => Timestamp -> Jvalue -> Maybe (MercadoBook p v) convertMercadoBook _ (Jobject dict) = do jBook <- Map.lookup "orderbook" dict case jBook of Jobject book -> do Jarray jasks <- Map.lookup "asks" book Jarray jbids <- Map.lookup "bids" book jLatestOid <- Map.lookup "latest_order_id" book oid' <- getNumber jLatestOid -- FIX ME! This fails due to lack of precision in 'Double' if OID > 2^52 oid <- maybePositive oid' as <- mapM makeMercadoAsk jasks bs <- mapM makeMercadoBid jbids return (QuoteBook { bids = sortBy (flip (compare `on` price)) bs --flip inverts order , asks = sortBy (compare `on` price) as , counter = OID 0 (round oid :: Word64) }) _ -> Nothing convertMercadoBook _ _ = Nothing ------------------------------------------------ type QuoteMaker p v = Jvalue -> Maybe (MercadoQuote p v) makeMercadoAsk :: (Fractional p, Fractional v) => QuoteMaker p v makeMercadoAsk = makeMercadoQuote toAsk makeMercadoBid :: (Fractional p, Fractional v) => QuoteMaker p v makeMercadoBid = makeMercadoQuote toBid type ToBidOrAsk p v = Double -> Double -> OrderID -> Bool -> MercadoQuote p v toBid :: (Fractional p, Fractional v) => Double -> Double -> OrderID -> Bool -> MercadoQuote p v toBid p v oid flag = Quote {side = Bid, price = realToFrac p, volume = realToFrac v, qtail=(oid, flag)} toAsk :: (Fractional p, Fractional v) => Double -> Double -> OrderID -> Bool -> MercadoQuote p v toAsk p v oid flag = Quote {side = Ask, price = realToFrac p, volume = realToFrac v, qtail=(oid, flag)} makeMercadoQuote :: ToBidOrAsk p v -> QuoteMaker p v makeMercadoQuote bidask (Jobject dict) = do jOid <- Map.lookup "order_id" dict oid' <- getNumber jOid -- FIX ME! This fails due to lack of precision in 'Double' if OID > 2^52 oid <- maybePositive oid' jPrice <- Map.lookup "limit_price" dict p' <- jStringToNumber jPrice p <- maybePositive p' jVol <- Map.lookup "quantity" dict v' <- jStringToNumber jVol v <- maybePositive v' jIsMine <- Map.lookup "is_owner" dict isMine <- getBool jIsMine return (bidask p v (OID 0 (round oid :: Word64)) isMine ) makeMercadoQuote _ _ = Nothing ------------------------------------------------------------------------------ convertLimitOrderResp :: (Fractional v, Fractional p) => Timestamp -> Jvalue -> Maybe (Order p v (Confirmation p v), [Fill p v]) convertLimitOrderResp _ (Jobject dict) = do jOrder <- Map.lookup "order" dict convertMercadoLimitOrder jOrder convertLimitOrderResp _ _ = Nothing ------------------------------------------------------------------------------ convertGetOrderResp :: (Fractional v, Fractional p) => Timestamp -> Jvalue -> Maybe [(Order p v (Confirmation p v), [Fill p v])] convertGetOrderResp _ (Jobject dict) = do jOrders <- Map.lookup "orders" dict case jOrders of Jarray jOrds -> mapM convertMercadoLimitOrder jOrds _ -> Nothing convertGetOrderResp _ _ = Nothing ------------------------------------------------------------------------------ -- | This functions converts the Jvalue representing a single Order. convertMercadoLimitOrder :: (Fractional v, Fractional p) => Jvalue -> Maybe (Order p v (Confirmation p v), [Fill p v]) convertMercadoLimitOrder (Jobject dict) = do jOid <- Map.lookup "order_id" dict jPair <- Map.lookup "coin_pair" dict jSide <- Map.lookup "order_type" dict jStatus <- Map.lookup "status" dict jHasFill<- Map.lookup "has_fills" dict jPrice <- Map.lookup "limit_price" dict jVolume <- Map.lookup "quantity" dict jExePri <- Map.lookup "executed_price_avg" dict jExeVol <- Map.lookup "executed_quantity" dict jFee <- Map.lookup "fee" dict jTime <- Map.lookup "created_timestamp" dict jUpdated<- Map.lookup "updated_timestamp" dict jFills <- Map.lookup "operations" dict case (jOid, jPair, jSide, jStatus, jHasFill, jPrice, jVolume, jExePri, jExeVol, jFee, jTime, jUpdated, jFills) of ( Jnumber orderID' , Jstring _coinPair , Jnumber nSide , Jnumber nStatus , Jbool hasFills , Jstring sPrice , Jstring sVolume , Jstring sExePri , Jstring sExeVol , Jstring _sFee -- ignored for now , Jstring sTime , Jstring _sUpdated -- ignored for now , Jarray jOps ) -> do let mStatus = case nStatus of 2 -> case hasFills of {False -> Just Active; True -> Just ActivePartiallyExecuted} 3 -> Just Inactive 4 -> Just Inactive _ -> Nothing price' <- (readMaybe sPrice :: Maybe Double) volume' <- (readMaybe sVolume :: Maybe Double) exePrice' <- (readMaybe sExePri :: Maybe Double) exeVolume' <- (readMaybe sExeVol :: Maybe Double) time' <- (readMaybe sTime :: Maybe Double) -- must be non-negative otherwise abort lOrderID <- maybePositive orderID' let oid = (OID 0 $ truncate lOrderID) price <- maybePositive price' volume <- maybePositive volume' exePrice <- maybePositive exePrice' exeVolume <- maybePositive exeVolume' time <- maybePositive time' fills <- mapM (convertMercadoOperation oid) jOps let requestedVolume = realToFrac volume executedVolume = realToFrac exeVolume mSide = case nSide of { 1 -> Just Bid; 2 -> Just Ask; _ -> Nothing } case (mSide,mStatus) of (_ , Nothing) -> Nothing -- invalid status field in response (Nothing , _) -> Nothing -- invalid order_type field in response (Just side, Just status) -> return ( LimitOrder { oSide = side , limitPrice = realToFrac price , limitVolume = requestedVolume , aConfirmation = Conf{ orderID = oid , mTimestamp = Just (truncate time) -- FIX ME! This should really be "updated_timestamp" (but this has other ramifications) , mOrderStatus = Just status , mExecuted = Just (realToFrac exePrice, executedVolume)} }, fills) _ -> Nothing convertMercadoLimitOrder _ = Nothing ------------------------------------------------------------------------------- -- | This functions converts the Jvalue representing a single Operation. convertMercadoOperation :: (Fractional p, Fractional v) => OrderID -> Jvalue -> Maybe (Fill p v) convertMercadoOperation oid (Jobject dict) = do jFID <- Map.lookup "operation_id" dict jPrice <- Map.lookup "price" dict jVol <- Map.lookup "quantity" dict jFeeRate <- Map.lookup "fee_rate" dict jTime <- Map.lookup "executed_timestamp" dict case (jFID,jPrice,jVol,jFeeRate,jTime) of (Jnumber fid', Jstring sPrice, Jstring sVol, Jstring sRate, Jstring sTime) -> do price' <- (readMaybe sPrice :: Maybe Double) vol' <- (readMaybe sVol :: Maybe Double) rate' <- (readMaybe sRate :: Maybe Double) time' <- (readMaybe sTime :: Maybe Double) price <- maybePositive price' vol <- maybePositive vol' rate <- maybePositive rate' time <- maybePositive time' return Fill { fillID = truncate fid' , fillPrice = realToFrac price , fillVolume = realToFrac vol , fillFee = realToFrac (0.01 * rate * price * vol) , mFillTime = Just (round time) , orderId = oid } _ -> Nothing convertMercadoOperation _ _ = Nothing ------------------------------------------------------------------------------ convertMercadoGetInfoResp :: Timestamp -> Jvalue -> Maybe (CurrencyVol BRL, Vol BTC, Vol LTC, Timestamp) convertMercadoGetInfoResp t (Jobject dict) = do jBalance <- Map.lookup "balance" dict jLimits <- Map.lookup "withdrawal_limits" dict case (jBalance,jLimits) of (Jobject balDict, Jobject _) -> do jReais <- Map.lookup "brl" balDict jBitcoins <- Map.lookup "btc" balDict jLTCs <- Map.lookup "ltc" balDict case (jReais, jBitcoins, jLTCs) of (Jobject realDict, Jobject btcDict, Jobject ltcDict) -> do jReaisAvail <- Map.lookup "available" realDict jBTCsAvail <- Map.lookup "available" btcDict jLTCsAvail <- Map.lookup "available" ltcDict case (jReaisAvail, jBTCsAvail, jLTCsAvail) of (Jstring sBRL, Jstring sBTC, Jstring sLTC) -> do brlFunds' <- (readMaybe sBRL :: Maybe Double) btcFunds' <- (readMaybe sBTC :: Maybe Double) ltcFunds' <- (readMaybe sLTC :: Maybe Double) brlFunds <- maybePositive brlFunds' -- must be non-negative btcFunds <- maybePositive btcFunds' -- otherwise abort ltcFunds <- maybePositive ltcFunds' -- otherwise abort return (realToFrac brlFunds, realToFrac btcFunds, realToFrac ltcFunds, t) _ -> Nothing _ -> Nothing _ -> Nothing convertMercadoGetInfoResp _ _ = Nothing ------------------------------------------------------------------------------ convertMercadoWithdrawalResponse :: (Fractional vol) => Timestamp -> Jvalue -> Maybe (MercadoTransfer vol) convertMercadoWithdrawalResponse _ (Jobject wDict) = do jWithdrawal <- Map.lookup "withdrawal" wDict case jWithdrawal of Jobject dict -> do jSt <- Map.lookup "status" dict jCoin <- Map.lookup "coin" dict jVol <- Map.lookup "quantity" dict jFee <- Map.lookup "fee" dict jWallet <- Map.lookup "address" dict jTime <- Map.lookup "created_timestamp" dict jId <- Map.lookup "id" dict _ <- Map.lookup "tx" dict _ <- Map.lookup "updated_timestamp" dict _ <- Map.lookup "description" dict case (jSt,jCoin,jVol,jFee,jWallet,jTime,jId) of (Jnumber st, Jstring _, Jstring sVol, Jstring sFee, Jstring sWallet, Jstring sTime, Jnumber tid') -> do if st /= 1 && st /= 2 then Nothing else do vol' <- (readMaybe sVol :: Maybe Double) vol <- maybePositive vol' fee' <- (readMaybe sFee :: Maybe Double) fee <- maybePositive fee' time'<- (readMaybe sTime :: Maybe Double) time <- maybePositive time' _ <- maybePositive tid' -- Mercado id not the bitcoin transaction id, not recording it. return MercadoTransfer { transfTime = truncate time , transfVol = realToFrac vol -- fee NOT included , transfFee = realToFrac fee -- this is also debited from account , toWallet = Wallet sWallet , mTransferID = Nothing } _ -> Nothing _ -> Nothing convertMercadoWithdrawalResponse _ _ = Nothing ------------------------------------------------------------------------------ convertMercadoTrades :: (Fractional p, Fractional v) => Jvalue -> Maybe [MercadoTrade p v] convertMercadoTrades (Jarray trades) = mapM convertMercadoTrade trades convertMercadoTrades _ = Nothing convertMercadoTrade :: (Fractional p, Fractional v) => Jvalue -> Maybe (MercadoTrade p v) convertMercadoTrade (Jobject dict) = do jTime <- Map.lookup "date" dict jPrice <- Map.lookup "price" dict jVol <- Map.lookup "amount" dict jTid <- Map.lookup "tid" dict jSide <- Map.lookup "type" dict case (jTime, jPrice, jVol, jTid, jSide) of (Jnumber time', Jnumber price', Jnumber vol', Jnumber tid', Jstring sSide) -> do time <- maybePositive time' price <- maybePositive price' vol <- maybePositive vol' tid <- maybePositive tid' side <- case sSide of "buy" -> Just Bid "sell" -> Just Ask _ -> Nothing return $ MT { tradeTime = truncate time , tradeSide = side , tradePrice = realToFrac price , tradeVolume = realToFrac vol , tradeId = truncate tid } _ -> Nothing convertMercadoTrade _ = Nothing
dimitri-xyz/mercado-bitcoin
src/Mercado/Internals.hs
bsd-3-clause
20,014
0
24
6,251
5,377
2,689
2,688
334
36
module Experiments.MonadT.ErrorIO where import Control.Monad.Except import System.Random getRandomNumber :: IO Int getRandomNumber = randomRIO (0, 20) data RandError = RandError Int String deriving (Show) type ErrorIO a = ExceptT RandError IO a errIfOdd :: ErrorIO Int errIfOdd = do n <- lift getRandomNumber if (n `mod` 2) == 1 then throwError $ RandError n "odd number error" else return n errIfOver :: Int -> ErrorIO Int errIfOver lim = do n <- lift getRandomNumber if n > lim then throwError $ RandError n "over limit" else return n errIfUnder :: Int -> ErrorIO Int errIfUnder lim = do n <- lift getRandomNumber if n < lim then throwError $ RandError n "under limit" else return n getSomeNumbers :: ErrorIO [Int] getSomeNumbers = do n1 <- errIfOdd n2 <- errIfOver 10 n3 <- errIfUnder 10 return [n1, n2, n3] getSomeNumbersEither :: IO (Either RandError [Int]) getSomeNumbersEither = runExceptT getSomeNumbers alwaysPass :: ErrorIO [Int] alwaysPass = do n1 <- errIfOver 30 n2 <- errIfUnder 0 n3 <- errIfOver 30 return [n1, n2, n3] alwaysPassEither :: IO (Either RandError [Int]) alwaysPassEither = runExceptT alwaysPass alwaysFail :: ErrorIO [Int] alwaysFail = do n1 <- errIfOver 30 n2 <- errIfUnder 0 n3 <- errIfOver 0 return [n1, n2, n3] alwaysFailEither :: IO (Either RandError [Int]) alwaysFailEither = runExceptT alwaysFail
rumblesan/haskell-experiments
src/Experiments/MonadT/ErrorIO.hs
bsd-3-clause
1,440
0
10
323
506
254
252
52
2
module Exercises109 where fibs :: Num a => [a] fibs = 1 : scanl (+) 1 fibs fibsN :: Num a => Int -> a fibsN x = fibs !! x -- 1. fibs20 = take 20 fibs -- 2. fibsUnder100 = takeWhile (< 100) fibs -- 3. fact :: (Num a, Enum a) => [a] fact = scanl (*) 1 [1..] factN x = fact !! x
pdmurray/haskell-book-ex
src/ch10/Exercises10.9.hs
bsd-3-clause
282
0
6
74
144
80
64
10
1
{-# LANGUAGE OverloadedStrings #-} module Security ( tokens , credential ) where import Web.Twitter.Conduit import Web.Authenticate.OAuth tokens :: OAuth tokens = twitterOAuth { oauthConsumerKey = "YOUR CONSUMER KEY" , oauthConsumerSecret = "YOUR CONSUMER SECRET" } credential :: Credential credential = Credential [ ("oauth_token", "YOUR ACCESS TOKEN") , ("oauth_token_secret", "YOUR ACCESS TOKEN SECRET") ]
jmct/PubBot
Security.hs
bsd-3-clause
458
0
7
102
80
51
29
14
1
{-# LANGUAGE TemplateHaskell #-} {-# LANGUAGE RankNTypes #-} {-# LANGUAGE KindSignatures #-} {-# LANGUAGE TypeFamilies #-} {-# LANGUAGE CPP #-} -- {-# OPTIONS -ddump-splices #-} import Language.Haskell.TH.ExpandSyns import Language.Haskell.TH import Language.Haskell.TH.Syntax import Util import Types main = do putStrLn "Basic test..." $(mkTest [t| forall a. Show a => a -> ForAll [] -> (Int,ApplyToInteger []) |] -- GHC 7.8 always seems to consider the body of 'ForallT' to have a 'PlainTV', -- whereas it always has a 'KindedTV' with GHC 7.10 (in both cases, it doesn't appear -- to matter whether the definition of 'ForAll' is actually written with a kind signature). #if MIN_VERSION_template_haskell(2,10,0) [t| forall a. Show a => a -> (forall (x :: *). [] x) -> (Int,[] Integer) |] #else [t| forall a. Show a => a -> (forall x. [] x) -> (Int,[] Integer) |] #endif ) putStrLn "Variable capture avoidance test..." $(let -- See comment about 'PlainTV'/'KindedTV' above #if MIN_VERSION_template_haskell(2,10,0) y_0 = KindedTV (mkName "y_0") StarT #else y_0 = PlainTV (mkName "y_0") #endif expectedExpansion = forallT [y_0] (cxt []) (conT ''Either `appT` varT' "y" `appT` varT' "y_0" --> conT ''Int) -- the naive (and wrong) result would be: -- forall y. (forall y. Either y y -> Int) in mkTest (forallT'' ["y"] (conT' "E" `appT` varT' "y")) (forallT'' ["y"] expectedExpansion)) putStrLn "Testing that it doesn't crash on type families (expanding them is not supported yet)" $(let t = [t| (DF1 Int, TF1 Int, AT1 Int) |] in mkTest t t) putStrLn "Testing that the args of type family applications are handled" $(mkTest [t| (DF1 Int', TF1 Int', AT1 Int') |] [t| (DF1 Int, TF1 Int, AT1 Int) |]) putStrLn "Higher-kinded synonym" $(mkTest [t| Either' (ListOf Int') (ListOf Char) |] [t| Either [Int] [Char] |]) putStrLn "Nested" $(mkTest [t| Int'' |] [t| Int |])
seereason/th-expand-syns
testing/Main.hs
bsd-3-clause
2,189
0
19
629
322
182
140
39
1
module Spec.Constant where import Data.Word(Word32, Word64) data Constant = Constant { cName :: String , cValueString :: String , cValue :: ConstantValue , cComment :: Maybe String } deriving (Show) data ConstantValue = -- | An integral value with no specific size IntegralValue Integer | -- | A value ending in 'f' FloatValue Float | -- A value sized with 'U' Word32Value Word32 | -- A value sized with 'ULL' Word64Value Word64 deriving (Show)
oldmanmike/vulkan
generate/src/Spec/Constant.hs
bsd-3-clause
693
0
9
322
97
60
37
13
0
-- -- Copyright (c) 2009-2011, ERICSSON AB -- All rights reserved. -- -- Redistribution and use in source and binary forms, with or without -- modification, are permitted provided that the following conditions are met: -- -- * Redistributions of source code must retain the above copyright notice, -- this list of conditions and the following disclaimer. -- * Redistributions in binary form must reproduce the above copyright -- notice, this list of conditions and the following disclaimer in the -- documentation and/or other materials provided with the distribution. -- * Neither the name of the ERICSSON AB nor the names of its contributors -- may be used to endorse or promote products derived from this software -- without specific prior written permission. -- -- THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" -- AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE -- IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE -- DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE -- FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL -- DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR -- SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER -- CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, -- OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -- OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. -- {-# LANGUAGE GADTs #-} {-# LANGUAGE ConstraintKinds #-} {-# LANGUAGE RecordWildCards #-} {-# LANGUAGE FlexibleContexts #-} -- | Module that translates from the UntypedFeld program from the middleend to -- the module format in the backend. module Feldspar.Compiler.Imperative.FromCore ( fromCore , fromCoreM , fromCoreExp , getCore' ) where import qualified Data.Map as Map import Data.Char (toLower) import Data.List (nub, find, isPrefixOf) import Data.Maybe (isJust, fromJust, mapMaybe) import Control.Monad.RWS import Control.Monad.State import Control.Applicative import Feldspar.Core.Types import Feldspar.Core.UntypedRepresentation ( Term(..), Lit(..), collectLetBinders, collectBinders , UntypedFeldF(App, LetFun), Fork(..) ) import qualified Feldspar.Core.UntypedRepresentation as Ut import Feldspar.Core.Middleend.FromTyped import Feldspar.Compiler.Backend.C.Platforms (extend, c99) import Feldspar.Core.Constructs (SyntacticFeld) import Feldspar.Core.Frontend (reifyFeldM) import qualified Feldspar.Compiler.Imperative.Representation as Rep (Variable(..), Type(..), ScalarType(..)) import Feldspar.Compiler.Imperative.Representation ( ActualParameter(..), Block(..), Declaration(..), Entity(..) , Expression(..), Module(..), Program(..), Pattern(..), ParType(..) , Constant(..), typeof, fv ) import Feldspar.Compiler.Imperative.Frontend import Feldspar.Compiler.Imperative.FromCore.Interpretation import Feldspar.Compiler.Backend.C.Options (Options(..)) import Feldspar.Compiler.Backend.C.MachineLowering {- Fast returns ------------ Fast returns really means single return value and that value fits in a register--thus they are platform dependent. This is why we have to compileTypeRep since that can make configuration specific choices about data layout. The user is free to ask for fast returns, but we might not be able to comply. For those cases we flip the option before generating code. -} -- | Get the generated core for a program with a specified output name. fromCore :: SyntacticFeld a => Options -> String -> a -> Module () fromCore opt funname prog = flip evalState 0 $ fromCoreM opt funname prog -- | Get the generated core for a program with a specified output name. fromCoreM :: (MonadState Integer m) => SyntacticFeld a => Options -> String -> a -> m (Module ()) fromCoreM opt funname prog = do s <- get let (ast, s') = flip runState (fromInteger s) $ reifyFeldM (frontendOpts opt) N32 prog uast = untype (frontendOpts opt) ast let opt' | useNativeReturns opt , not $ canFastReturn $ compileTypeRep opt (typeof uast) = opt { useNativeReturns = False } -- Note [Fast returns] | otherwise = opt fastRet = useNativeReturns opt' let (outParam,States s'',results) = runRWS (compileProgTop opt' uast) (initReader opt) $ States $ toInteger s' put s'' let decls = decl results ins = params results post = epilogue results ++ returns Block ds p = block results outDecl = Declaration outParam Nothing paramTypes = getTypes $ outDecl:map (`Declaration` Nothing) ins defs = nub (def results ++ paramTypes) ++ topProc (outs, ds', returns) | fastRet = ( Right outParam, outDecl:ds ++ decls , [call "return" [ValueParameter $ varToExpr outParam]]) | otherwise = ( Left [outParam], ds ++ decls, []) topProc = [Proc funname False ins outs $ Just (Block ds' (Sequence mainProg))] mainProg | Just _ <- find isTask $ def results = call "taskpool_init" [four,four,four] : p : call "taskpool_shutdown" [] : post | otherwise = p:post isTask (Proc{..}) = isPrefixOf "task_core" procName isTask _ = False four = ValueParameter $ ConstExpr $ IntConst 4 $ Rep.NumType Ut.Unsigned Ut.S32 return $ Module defs -- | Get the generated core for a program and an expression that contains the output. The components -- of the result are as follows, in order: -- -- * A list of extra entities needed by the program -- * A list of declarations needed by the program -- * The actual program -- * An expression that contains the result -- * A list of epilogue programs, for freeing memory, etc. fromCoreExp :: (MonadState Integer m) => SyntacticFeld a => Options -> Map.Map Integer String -> a -> m ([Entity ()], [Declaration ()], Program (), Expression (), [Program ()]) fromCoreExp opt aliases prog = do s <- get let (ast, s') = flip runState (fromInteger s) $ reifyFeldM (frontendOpts opt) N32 prog uast = untype (frontendOpts opt) ast mkAlias (Ut.Var i t) = do n <- Map.lookup (toInteger i) aliases return (i, varToExpr $ Rep.Variable (compileTypeRep opt t) n) as = mapMaybe mkAlias $ Ut.fv uast let (expr,States s'',results) = runRWS (compileExpr (CEnv opt False) uast) (Readers as opt) $ States $ toInteger s' put s'' unless (null (params results)) $ error "fromCoreExp: unexpected params" let x = getPlatformRenames opt Block ls p = block results return ( renameEnt opt x <$> def results , renameDecl x <$> (ls ++ decl results) , renameProg opt x p , renameExp x expr , renameProg opt x <$> epilogue results ) -- | Get the generated core for a program. getCore' :: SyntacticFeld a => Options -> a -> Module () getCore' opts = fromCore opts "test" compileProgTop :: Options -> Ut.UntypedFeld -> CodeWriter (Rep.Variable ()) compileProgTop opt (In (Ut.Lambda (Ut.Var v ta) body)) = do let typ = compileTypeRep opt ta (arg,arge) | Rep.StructType{} <- typ = (mkPointer typ v, Deref $ varToExpr arg) | otherwise = (mkVariable typ v, varToExpr arg) tell $ mempty {params=[arg]} withAlias v arge $ compileProgTop opt body compileProgTop opt (In (Ut.App Ut.Let _ [In (Ut.Literal l), In (Ut.Lambda (Ut.Var v _) body)])) | representableType l = do tellDef [ValueDef var c] withAlias v (varToExpr var) $ compileProgTop opt body where var = mkVariable (typeof c) v -- Note [Precise size information] c = literalConst opt l compileProgTop opt a = do let outType' = compileTypeRep opt (typeof a) (outType, outLoc) | useNativeReturns opt = (outType', varToExpr outParam) | otherwise = (Rep.MachineVector 1 (Rep.Pointer outType'), Deref $ varToExpr outParam) outParam = Rep.Variable outType "out" compileProg (cenv0 opt) (Just outLoc) a return outParam {- Precise size information ------------------------ Tight size bounds for a given literal is easy to compute. Precise bounds are particularly important for array literals since they are often copied in the deepCopy function near the CodeGen in the backend. Deepcopy will appear to hang when generating the copy code for insanely large array literals, so don't do that. -} data CompileEnv = CEnv { opts :: Options , inTask :: Bool } -- | Initial environment for compile. cenv0 :: Options -> CompileEnv cenv0 opts = CEnv opts False -- | Compiles code and assigns the expression to the given location. compileExprLoc :: CompileEnv -> Location -> Ut.UntypedFeld -> CodeWriter () compileExprLoc env loc e = do expr <- compileExpr env e assign loc expr -- | Compiles code into a fresh variable. compileProgFresh :: CompileEnv -> Ut.UntypedFeld -> CodeWriter (Expression ()) compileProgFresh env e = do loc <- freshVar (opts env) "e" (typeof e) compileProg env (Just loc) e return loc -- | Compile an expression and make sure that the result is stored in a variable compileExprVar :: CompileEnv -> Ut.UntypedFeld -> CodeWriter (Expression ()) compileExprVar env e = do e' <- compileExpr env e case e' of _ | isNearlyVar e' -> return e' _ -> do varId <- freshId let loc = varToExpr $ mkNamedVar "e" (typeof e') varId declare loc assign (Just loc) e' return loc where isNearlyVar VarExpr{} = True isNearlyVar (Deref a) = isNearlyVar a isNearlyVar (AddrOf a) = isNearlyVar a isNearlyVar _ = False -- | Compile a function bound by a LetFun. compileFunction :: CompileEnv -> Expression () -> (String, Fork, Ut.UntypedFeld) -> CodeWriter () compileFunction env loc (coreName, kind, e) | (bs, e') <- collectBinders e = do es' <- mapM (compileExpr env) (map (In . Ut.Variable) bs) let args = nub $ map exprToVar es' ++ fv loc -- Task core: ((_, ws), Block ds bl) <- confiscateBigBlock $ case kind of Future -> do p' <- compileExprVar env {inTask = True } e' tellProg [iVarPut loc p'] Par -> compileProg env {inTask = True} (Just loc) e' Loop | (_:_) <- es' , Ut.ElementsType{} <- typeof e' -> compileProg env (Just loc) e' | (ix:_) <- es' -> compileProg env (Just $ ArrayElem loc ix) e' None -> compileProg env (Just loc) e' tellDef [Proc coreName (kind == Loop) args (Left []) $ Just $ Block (decl ws ++ ds) bl] -- Task: let taskName = "task" ++ drop 9 coreName runTask = Just $ toBlock $ run coreName args outs = [mkNamedRef "params" Rep.VoidType (-1)] case kind of _ | kind `elem` [None, Loop] -> return () _ -> tellDef [Proc taskName False [] (Left outs) runTask] -- | Create a variable of the right type for storing a length. mkLength :: CompileEnv -> Ut.UntypedFeld -> Ut.Type -> CodeWriter (Expression ()) mkLength env a t | isVariableOrLiteral a = compileExpr env a | otherwise = do lenvar <- freshVar (opts env) "len" t compileProg env (Just lenvar) a return lenvar mkBranch :: CompileEnv -> Location -> Ut.UntypedFeld -> Ut.UntypedFeld -> Maybe Ut.UntypedFeld -> CodeWriter () mkBranch env loc c th el = do ce <- compileExpr env c (_, tb) <- confiscateBlock $ compileProg env loc th (_, eb) <- if isJust el then confiscateBlock $ compileProg env loc (fromJust el) else return (undefined, toBlock Empty) tellProg [Switch ce [(Pat (litB True), tb), (Pat (litB False), eb)]] compileProg :: CompileEnv -> Location -> Ut.UntypedFeld -> CodeWriter () -- Array compileProg env (Just loc) (In (App Ut.Parallel _ [len, In (Ut.Lambda (Ut.Var v ta) ixf)])) = do let ix = mkVar (compileTypeRep (opts env) ta) v len' <- mkLength env len ta (ptyp, b) <- case ixf of In (App (Ut.Call Loop n) _ vs) -> do vs' <- mapM (compileExpr env) vs let mkV w = Rep.Variable (typeof w) (lName w) args = map (ValueParameter . varToExpr) $ nub $ map mkV vs' ++ fv loc return $ (TaskParallel, toBlock $ ProcedureCall n args) _ -> do b' <- confiscateBlock $ compileProg env (Just $ ArrayElem loc ix) ixf return (Parallel, snd b') tellProg [initArray (Just loc) len'] tellProg [for ptyp (lName ix) (litI32 0) len' (litI32 1) b] compileProg env loc (In (App Ut.Sequential _ [len, init', In (Ut.Lambda (Ut.Var v tix) ixf1)])) | In (Ut.Lambda (Ut.Var s tst) l) <- ixf1 , (bs, In (Ut.App Ut.Tup2 _ [In (Ut.Variable t1), In (Ut.Variable t2)])) <- collectLetBinders l , not $ null bs , (e, step) <- last bs , t1 == e , t2 == e = do blocks <- mapM (confiscateBlock . compileBind env) (init bs) let (dss, lets) = unzip $ map (\(_, Block ds (Sequence body)) -> (ds, body)) blocks let ix = mkVar (compileTypeRep (opts env) tix) v len' <- mkLength env len tix st1 <- freshVar (opts env) "st" tst let st = mkRef (compileTypeRep (opts env) tst) s st_val = Deref st declareAlias st (_, Block ds (Sequence body)) <- confiscateBlock $ withAlias s st_val $ compileProg env (ArrayElem <$> loc <*> pure ix) step withAlias s st_val $ compileProg env (Just st1) init' tellProg [ Assign (Just st) (AddrOf st1) , initArray loc len'] tellProg [toProg $ Block (concat dss ++ ds) $ for Sequential (lName ix) (litI32 0) len' (litI32 1) $ toBlock $ Sequence (concat lets ++ body ++ maybe [] (\arr -> [Assign (Just st) $ AddrOf (ArrayElem arr ix)]) loc)] compileProg env loc (In (App Ut.Sequential _ [len, st, In (Ut.Lambda (Ut.Var v t) (In (Ut.Lambda (Ut.Var s _) step)))])) = do let tr' = typeof step let ix = mkVar (compileTypeRep (opts env) t) v len' <- mkLength env len t tmp <- freshVar (opts env) "seq" tr' (_, Block ds (Sequence body)) <- confiscateBlock $ withAlias s (StructField tmp "member2") $ compileProg env (Just tmp) step tellProg [initArray loc len'] compileProg env (Just $ StructField tmp "member2") st tellProg [toProg $ Block ds $ for Sequential (lName ix) (litI32 0) len' (litI32 1) $ toBlock $ Sequence $ body ++ [copyProg (ArrayElem <$> loc <*> pure ix) [StructField tmp "member1"] ]] compileProg env loc (In (App Ut.Append _ [a, b])) = do a' <- compileExpr env a b' <- compileExpr env b tellProg [copyProg loc [a', b']] compileProg env loc (In (App Ut.SetIx _ [arr, i, a])) = do compileProg env loc arr i' <- compileExpr env i compileProg env (ArrayElem <$> loc <*> pure i') a compileProg env (Just loc) (In (App Ut.GetIx _ [arr, i])) = do a' <- compileExpr env arr i' <- compileExpr env i let el = ArrayElem a' i' tellProg $ if isArray $ typeof el then [Assign (Just loc) el] else [copyProg (Just loc) [el]] compileProg env loc (In (App Ut.SetLength _ [len, arr])) = do len' <- compileExpr env len compileProg env loc arr tellProg [setLength loc len'] -- Binding compileProg _ _ e@(In Ut.Lambda{}) = error ("Can only compile top-level lambda: " ++ show e) compileProg env loc (In (Ut.App Ut.Let _ [a, In (Ut.Lambda (Ut.Var v ta) body)])) = do e <- compileLet env a ta v withAlias v e $ compileProg env loc body -- Bits -- Complex -- Condition compileProg env loc (In (App Ut.Condition _ [cond, tHEN, eLSE])) = mkBranch env loc cond tHEN $ Just eLSE compileProg env loc (In (App Ut.ConditionM _ [cond, tHEN, eLSE])) = mkBranch env loc cond tHEN $ Just eLSE -- Conversion -- Elements compileProg env loc (In (App Ut.EMaterialize _ [len, arr])) = do len' <- mkLength env len (Ut.typeof len) tellProg [initArray loc len'] compileProg env loc arr compileProg env (Just loc) (In (App Ut.EWrite _ [ix, e])) = do dst <- compileExpr env ix compileProg env (Just $ ArrayElem loc dst) e compileProg _ _ (In (App Ut.ESkip _ _)) = return () compileProg env loc (In (App Ut.EPar _ [p1, p2])) = do (_, Block ds1 b1) <- confiscateBlock $ compileProg env loc p1 (_, Block ds2 b2) <- confiscateBlock $ compileProg env loc p2 tellProg [toProg $ Block (ds1 ++ ds2) (Sequence [b1,b2])] compileProg env (Just loc) (In (App Ut.EparFor _ [len, In (Ut.Lambda (Ut.Var v ta) ixf)])) = do let ix = mkVar (compileTypeRep (opts env) ta) v len' <- mkLength env len ta (ptyp, b) <- case ixf of In (App (Ut.Call Loop n) _ vs) -> do vs' <- mapM (compileExpr env) vs let mkV w = Rep.Variable (typeof w) (lName w) args = map (ValueParameter . varToExpr) $ nub $ map mkV vs' ++ fv loc return $ (TaskParallel, toBlock $ ProcedureCall n args) _ -> do b' <- confiscateBlock $ compileProg env (Just loc) ixf return (Parallel, snd b') tellProg [for ptyp (lName ix) (litI32 0) len' (litI32 1) b] -- Error compileProg _ _ (In (App Ut.Undefined _ _)) = return () compileProg env loc (In (App (Ut.Assert msg) _ [cond, a])) = do compileAssert env cond msg compileProg env loc a -- Future compileProg _ _ e@(In (App Ut.MkFuture _ _)) = error ("Unexpected MkFuture:" ++ show e) compileProg env (Just loc) (In (LetFun f e)) = do compileFunction env loc f compileProg env (Just loc) e compileProg env loc (In (App Ut.Await _ [a])) = do fut <- compileExprVar env a tellProg [iVarGet (inTask env) l fut | Just l <- [loc]] -- Literal compileProg env loc (In (Ut.Literal a)) = case loc of Just l -> literalLoc (opts env) l a Nothing -> return () -- Logic -- Loop compileProg env (Just loc) (In (App Ut.ForLoop _ [len, init', In (Ut.Lambda (Ut.Var ix ta) (In (Ut.Lambda (Ut.Var st _) ixf)))])) = do let ix' = mkVar (compileTypeRep (opts env) ta) ix len' <- mkLength env len ta (lstate, stvar) <- mkDoubleBufferState loc st compileProg env (Just lstate) init' (_, Block ds body) <- withAlias st lstate $ confiscateBlock $ compileProg env (Just stvar) ixf >> shallowCopyWithRefSwap lstate stvar tellProg [toProg $ Block ds (for Sequential (lName ix') (litI32 0) len' (litI32 1) (toBlock body))] shallowAssign (Just loc) lstate compileProg env (Just loc) (In (App Ut.WhileLoop _ [init', In (Ut.Lambda (Ut.Var cv _) cond), In (Ut.Lambda (Ut.Var bv _) body)])) = do let condv = mkVar (compileTypeRep (opts env) (typeof cond)) cv (lstate,stvar) <- mkDoubleBufferState loc bv compileProg env (Just lstate) init' (_, cond') <- confiscateBlock $ withAlias cv lstate $ compileProg env (Just condv) cond (_, body') <- withAlias bv lstate $ confiscateBlock $ compileProg env (Just stvar) body >> shallowCopyWithRefSwap lstate stvar declare condv tellProg [while cond' condv body'] shallowAssign (Just loc) lstate -- LoopM compileProg env loc (In (App Ut.While _ [cond,step])) = do condv <- freshVar (opts env) "cond" (typeof cond) (_, cond') <- confiscateBlock $ compileProg env (Just condv) cond (_, step') <- confiscateBlock $ compileProg env loc step tellProg [while cond' condv step'] compileProg env loc (In (App Ut.For _ [len, In (Ut.Lambda (Ut.Var v ta) ixf)])) = do let ix = mkVar (compileTypeRep (opts env) ta) v len' <- mkLength env len ta (_, Block ds body) <- confiscateBlock $ compileProg env loc ixf tellProg [toProg $ Block ds (for Sequential (lName ix) (litI32 0) len' (litI32 1) (toBlock body))] -- Mutable compileProg env loc (In (App Ut.Run _ [ma])) = compileProg env loc ma compileProg env loc (In (App Ut.Return t [a])) | Ut.MutType Ut.UnitType <- t = return () | Ut.ParType Ut.UnitType <- t = return () | otherwise = compileProg env loc a compileProg env loc (In (App Ut.Bind _ [ma, In (Ut.Lambda (Ut.Var v ta) body)])) | (In (App Ut.ParNew _ _)) <- ma = do let var = mkVar (compileTypeRep (opts env) ta) v declare var tellProg [iVarInit (AddrOf var)] compileProg env loc body | otherwise = do let var = mkVar (compileTypeRep (opts env) ta) v declare var compileProg env (Just var) ma compileProg env loc body compileProg env loc (In (App Ut.Then _ [ma, mb])) = do compileProg env Nothing ma compileProg env loc mb compileProg env loc (In (App Ut.When _ [c, action])) = mkBranch env loc c action Nothing -- MutableArray compileProg env loc (In (App Ut.NewArr _ [len, a])) = do nId <- freshId let var = mkNamedVar "i" (Rep.MachineVector 1 (Rep.NumType Ut.Unsigned Ut.S32)) nId ix = varToExpr var a' <- compileExpr env a l <- compileExpr env len tellProg [initArray loc l] tellProg [for Sequential (Rep.varName var) (litI32 0) l (litI32 1) $ toBlock (Sequence [copyProg (ArrayElem <$> loc <*> pure ix) [a']])] compileProg env loc (In (App Ut.NewArr_ _ [len])) = do l <- compileExpr env len tellProg [initArray loc l] compileProg env loc (In (App Ut.GetArr _ [arr, i])) = do arr' <- compileExpr env arr i' <- compileExpr env i assign loc (ArrayElem arr' i') compileProg env _ (In (App Ut.SetArr _ [arr, i, a])) = do arr' <- compileExpr env arr i' <- compileExpr env i a' <- compileExpr env a assign (Just $ ArrayElem arr' i') a' -- MutableReference compileProg env loc (In (App Ut.NewRef _ [a])) = compileProg env loc a compileProg env loc (In (App Ut.GetRef _ [r])) = compileProg env loc r compileProg env _ (In (App Ut.SetRef _ [r, a])) = do var <- compileExpr env r compileProg env (Just var) a compileProg env _ (In (App Ut.ModRef _ [r, In (Ut.Lambda (Ut.Var v _) body)])) = do var <- compileExpr env r withAlias v var $ compileProg env (Just var) body -- Since the modifier function is pure it is safe to alias -- v with var here -- MutableToPure compileProg env (Just loc) (In (App Ut.RunMutableArray _ [marr])) | (In (App Ut.Bind _ [In (App Ut.NewArr_ _ [l]), In (Ut.Lambda (Ut.Var v _) body)])) <- marr , (In (App Ut.Return _ [In (Ut.Variable (Ut.Var r _))])) <- chaseBind body , v == r = do len <- compileExpr env l tellProg [setLength (Just loc) len] withAlias v loc $ compileProg env (Just loc) body compileProg env loc (In (App Ut.RunMutableArray _ [marr])) = compileProg env loc marr compileProg env loc (In (App Ut.WithArray _ [marr@(In Ut.Variable{}), In (Ut.Lambda (Ut.Var v _) body)])) = do e <- compileExpr env marr withAlias v e $ do b <- compileExpr env body tellProg [copyProg loc [b]] compileProg env loc (In (App Ut.WithArray _ [marr, In (Ut.Lambda (Ut.Var v ta) body)])) = do let var = mkVar (compileTypeRep (opts env) ta) v declare var compileProg env (Just var) marr e <- compileExpr env body tellProg [copyProg loc [e]] -- Noinline compileProg _ (Just _) (In (App Ut.NoInline _ [e])) = error ("Unexpected NoInline:" ++ show e) -- Par compileProg env loc (In (App Ut.ParRun _ [p])) = compileProg env loc p compileProg _ _ (In (App Ut.ParNew _ _)) = return () compileProg env loc (In (App Ut.ParGet _ [r])) = do iv <- compileExpr env r tellProg [iVarGet (inTask env) l iv | Just l <- [loc]] compileProg env _ (In (App Ut.ParPut _ [r, a])) = do iv <- compileExpr env r val <- compileExpr env a i <- freshId let var = varToExpr $ mkNamedVar "msg" (typeof val) i declare var assign (Just var) val tellProg [iVarPut iv var] compileProg _ _ (In (App Ut.ParFork _ [e])) = error ("Unexpected ParFork:" ++ show e) compileProg _ _ (In (App Ut.ParYield _ _)) = return () -- SizeProp compileProg env loc (In (App Ut.PropSize _ [e])) = compileProg env loc e -- SourceInfo compileProg env loc (In (App (Ut.SourceInfo info) _ [a])) = do tellProg [Comment True info] compileProg env loc a -- Switch compileProg env loc (In (App Ut.Switch _ [tree@(In (App Ut.Condition _ [In (App Ut.Equal _ [_, s]), _, _]))])) = do scrutinee <- compileExpr env s alts <- chaseTree env loc s tree tellProg [Switch{..}] compileProg env loc (In (App Ut.Switch _ [tree])) = compileProg env loc tree -- Tuple compileProg env loc (In (App Ut.Tup2 _ [m1, m2])) = do compileProg env (StructField <$> loc <*> pure "member1") m1 compileProg env (StructField <$> loc <*> pure "member2") m2 compileProg env loc (In (App Ut.Tup3 _ [m1, m2, m3])) = do compileProg env (StructField <$> loc <*> pure "member1") m1 compileProg env (StructField <$> loc <*> pure "member2") m2 compileProg env (StructField <$> loc <*> pure "member3") m3 compileProg env loc (In (App Ut.Tup4 _ [m1, m2, m3, m4])) = do compileProg env (StructField <$> loc <*> pure "member1") m1 compileProg env (StructField <$> loc <*> pure "member2") m2 compileProg env (StructField <$> loc <*> pure "member3") m3 compileProg env (StructField <$> loc <*> pure "member4") m4 compileProg env loc (In (App Ut.Tup5 _ [m1, m2, m3, m4, m5])) = do compileProg env (StructField <$> loc <*> pure "member1") m1 compileProg env (StructField <$> loc <*> pure "member2") m2 compileProg env (StructField <$> loc <*> pure "member3") m3 compileProg env (StructField <$> loc <*> pure "member4") m4 compileProg env (StructField <$> loc <*> pure "member5") m5 compileProg env loc (In (App Ut.Tup6 _ [m1, m2, m3, m4, m5, m6])) = do compileProg env (StructField <$> loc <*> pure "member1") m1 compileProg env (StructField <$> loc <*> pure "member2") m2 compileProg env (StructField <$> loc <*> pure "member3") m3 compileProg env (StructField <$> loc <*> pure "member4") m4 compileProg env (StructField <$> loc <*> pure "member5") m5 compileProg env (StructField <$> loc <*> pure "member6") m6 compileProg env loc (In (App Ut.Tup7 _ [m1, m2, m3, m4, m5, m6, m7])) = do compileProg env (StructField <$> loc <*> pure "member1") m1 compileProg env (StructField <$> loc <*> pure "member2") m2 compileProg env (StructField <$> loc <*> pure "member3") m3 compileProg env (StructField <$> loc <*> pure "member4") m4 compileProg env (StructField <$> loc <*> pure "member5") m5 compileProg env (StructField <$> loc <*> pure "member6") m6 compileProg env (StructField <$> loc <*> pure "member7") m7 compileProg env loc (In (App Ut.Tup8 _ [m1, m2, m3, m4, m5, m6, m7, m8])) = do compileProg env (StructField <$> loc <*> pure "member1") m1 compileProg env (StructField <$> loc <*> pure "member2") m2 compileProg env (StructField <$> loc <*> pure "member3") m3 compileProg env (StructField <$> loc <*> pure "member4") m4 compileProg env (StructField <$> loc <*> pure "member5") m5 compileProg env (StructField <$> loc <*> pure "member6") m6 compileProg env (StructField <$> loc <*> pure "member7") m7 compileProg env (StructField <$> loc <*> pure "member8") m8 compileProg env loc (In (App Ut.Tup9 _ [m1, m2, m3, m4, m5, m6, m7, m8, m9])) = do compileProg env (StructField <$> loc <*> pure "member1") m1 compileProg env (StructField <$> loc <*> pure "member2") m2 compileProg env (StructField <$> loc <*> pure "member3") m3 compileProg env (StructField <$> loc <*> pure "member4") m4 compileProg env (StructField <$> loc <*> pure "member5") m5 compileProg env (StructField <$> loc <*> pure "member6") m6 compileProg env (StructField <$> loc <*> pure "member7") m7 compileProg env (StructField <$> loc <*> pure "member8") m8 compileProg env (StructField <$> loc <*> pure "member9") m9 compileProg env loc (In (App Ut.Tup10 _ [m1, m2, m3, m4, m5, m6, m7, m8, m9, m10])) = do compileProg env (StructField <$> loc <*> pure "member1") m1 compileProg env (StructField <$> loc <*> pure "member2") m2 compileProg env (StructField <$> loc <*> pure "member3") m3 compileProg env (StructField <$> loc <*> pure "member4") m4 compileProg env (StructField <$> loc <*> pure "member5") m5 compileProg env (StructField <$> loc <*> pure "member6") m6 compileProg env (StructField <$> loc <*> pure "member7") m7 compileProg env (StructField <$> loc <*> pure "member8") m8 compileProg env (StructField <$> loc <*> pure "member9") m9 compileProg env (StructField <$> loc <*> pure "member10") m10 compileProg env loc (In (App Ut.Tup11 _ [m1, m2, m3, m4, m5, m6, m7, m8, m9, m10, m11])) = do compileProg env (StructField <$> loc <*> pure "member1") m1 compileProg env (StructField <$> loc <*> pure "member2") m2 compileProg env (StructField <$> loc <*> pure "member3") m3 compileProg env (StructField <$> loc <*> pure "member4") m4 compileProg env (StructField <$> loc <*> pure "member5") m5 compileProg env (StructField <$> loc <*> pure "member6") m6 compileProg env (StructField <$> loc <*> pure "member7") m7 compileProg env (StructField <$> loc <*> pure "member8") m8 compileProg env (StructField <$> loc <*> pure "member9") m9 compileProg env (StructField <$> loc <*> pure "member10") m10 compileProg env (StructField <$> loc <*> pure "member11") m11 compileProg env loc (In (App Ut.Tup12 _ [m1, m2, m3, m4, m5, m6, m7, m8, m9, m10, m11, m12])) = do compileProg env (StructField <$> loc <*> pure "member1") m1 compileProg env (StructField <$> loc <*> pure "member2") m2 compileProg env (StructField <$> loc <*> pure "member3") m3 compileProg env (StructField <$> loc <*> pure "member4") m4 compileProg env (StructField <$> loc <*> pure "member5") m5 compileProg env (StructField <$> loc <*> pure "member6") m6 compileProg env (StructField <$> loc <*> pure "member7") m7 compileProg env (StructField <$> loc <*> pure "member8") m8 compileProg env (StructField <$> loc <*> pure "member9") m9 compileProg env (StructField <$> loc <*> pure "member10") m10 compileProg env (StructField <$> loc <*> pure "member11") m11 compileProg env (StructField <$> loc <*> pure "member12") m12 compileProg env loc (In (App Ut.Tup13 _ [m1, m2, m3, m4, m5, m6, m7, m8, m9, m10, m11, m12, m13])) = do compileProg env (StructField <$> loc <*> pure "member1") m1 compileProg env (StructField <$> loc <*> pure "member2") m2 compileProg env (StructField <$> loc <*> pure "member3") m3 compileProg env (StructField <$> loc <*> pure "member4") m4 compileProg env (StructField <$> loc <*> pure "member5") m5 compileProg env (StructField <$> loc <*> pure "member6") m6 compileProg env (StructField <$> loc <*> pure "member7") m7 compileProg env (StructField <$> loc <*> pure "member8") m8 compileProg env (StructField <$> loc <*> pure "member9") m9 compileProg env (StructField <$> loc <*> pure "member10") m10 compileProg env (StructField <$> loc <*> pure "member11") m11 compileProg env (StructField <$> loc <*> pure "member12") m12 compileProg env (StructField <$> loc <*> pure "member13") m13 compileProg env loc (In (App Ut.Tup14 _ [m1, m2, m3, m4, m5, m6, m7, m8, m9, m10, m11, m12, m13, m14])) = do compileProg env (StructField <$> loc <*> pure "member1") m1 compileProg env (StructField <$> loc <*> pure "member2") m2 compileProg env (StructField <$> loc <*> pure "member3") m3 compileProg env (StructField <$> loc <*> pure "member4") m4 compileProg env (StructField <$> loc <*> pure "member5") m5 compileProg env (StructField <$> loc <*> pure "member6") m6 compileProg env (StructField <$> loc <*> pure "member7") m7 compileProg env (StructField <$> loc <*> pure "member8") m8 compileProg env (StructField <$> loc <*> pure "member9") m9 compileProg env (StructField <$> loc <*> pure "member10") m10 compileProg env (StructField <$> loc <*> pure "member11") m11 compileProg env (StructField <$> loc <*> pure "member12") m12 compileProg env (StructField <$> loc <*> pure "member13") m13 compileProg env (StructField <$> loc <*> pure "member14") m14 compileProg env loc (In (App Ut.Tup15 _ [m1, m2, m3, m4, m5, m6, m7, m8, m9, m10, m11, m12, m13, m14, m15])) = do compileProg env (StructField <$> loc <*> pure "member1") m1 compileProg env (StructField <$> loc <*> pure "member2") m2 compileProg env (StructField <$> loc <*> pure "member3") m3 compileProg env (StructField <$> loc <*> pure "member4") m4 compileProg env (StructField <$> loc <*> pure "member5") m5 compileProg env (StructField <$> loc <*> pure "member6") m6 compileProg env (StructField <$> loc <*> pure "member7") m7 compileProg env (StructField <$> loc <*> pure "member8") m8 compileProg env (StructField <$> loc <*> pure "member9") m9 compileProg env (StructField <$> loc <*> pure "member10") m10 compileProg env (StructField <$> loc <*> pure "member11") m11 compileProg env (StructField <$> loc <*> pure "member12") m12 compileProg env (StructField <$> loc <*> pure "member13") m13 compileProg env (StructField <$> loc <*> pure "member14") m14 compileProg env (StructField <$> loc <*> pure "member15") m15 -- Special case foreign imports since they can be of void type and just have effects. compileProg env loc (In (App p@Ut.ForeignImport{} t es)) = do es' <- mapM (compileExpr env) es tellProg [Assign loc $ fun' (compileTypeRep (opts env) t) (compileOp p) es'] -- Common nodes compileProg env (Just loc) (In (App (Ut.Call f name) _ es)) = do es' <- mapM (compileExpr env) es let args = nub $ map exprToVar es' ++ fv loc tellProg [iVarInitCond f (AddrOf loc)] tellProg [spawn f name args] compileProg env loc e = compileExprLoc env loc e compileExpr :: CompileEnv -> Ut.UntypedFeld -> CodeWriter (Expression ()) -- Array compileExpr env (In (App Ut.GetLength _ [a])) = do aExpr <- compileExpr env a return $ arrayLength aExpr compileExpr env (In (App Ut.GetIx _ [arr, i])) = do a' <- compileExpr env arr i' <- compileExpr env i return $ ArrayElem a' i' -- Bits compileExpr env (In (App Ut.Bit t [arr])) = do a' <- compileExpr env arr let t' = compileTypeRep (opts env) t return $ binop t' "<<" (litI t' 1) a' -- Binding compileExpr env (In (Ut.Variable (Ut.Var v t))) = do env' <- ask case lookup v (alias env') of Nothing -> return $ mkVar (compileTypeRep (opts env) t) v Just e -> return e compileExpr env (In (Ut.App Ut.Let _ [a, In (Ut.Lambda (Ut.Var v ta) body)])) = do e <- compileLet env a ta v withAlias v e $ compileExpr env body -- Bits -- Condition -- Conversion compileExpr env (In (App Ut.F2I t es)) = do es' <- mapM (compileExpr env) es let f' = fun (Rep.MachineVector 1 Rep.FloatType) "truncf" es' return $ Cast (compileTypeRep (opts env) t) f' compileExpr env (In (App Ut.I2N t1 [e])) | (Rep.MachineVector 1 (Rep.ComplexType t)) <- t' = do e' <- compileExpr env e let args = [Cast t e', litF 0] return $ fun t' (extend c99 "complex" t) args | otherwise = do e' <- compileExpr env e return $ Cast t' e' where t' = compileTypeRep (opts env) t1 compileExpr env (In (App Ut.B2I t [e])) = do e' <- compileExpr env e return $ Cast (compileTypeRep (opts env) t) e' compileExpr env (In (App Ut.Round t es)) = do es' <- mapM (compileExpr env) es let f' = fun (Rep.MachineVector 1 Rep.FloatType) "roundf" es' return $ Cast (compileTypeRep (opts env) t) f' compileExpr env (In (App Ut.Ceiling t es)) = do es' <- mapM (compileExpr env) es let f' = fun (Rep.MachineVector 1 Rep.FloatType) "ceilf" es' return $ Cast (compileTypeRep (opts env) t) f' compileExpr env (In (App Ut.Floor t es)) = do es' <- mapM (compileExpr env) es let f' = fun (Rep.MachineVector 1 Rep.FloatType) "floorf" es' return $ Cast (compileTypeRep (opts env) t) f' -- Error compileExpr env (In (App (Ut.Assert msg) _ [cond, a])) = do compileAssert env cond msg compileExpr env a -- Eq -- FFI -- Floating compileExpr _ (In (App Ut.Pi _ [])) = error "No pi ready" -- Fractional -- Future -- Literal compileExpr env (In (Ut.Literal l)) = literal (opts env) l -- Loop -- Logic -- Mutable compileExpr env (In (App Ut.Run _ [ma])) = compileExpr env ma -- MutableArray compileExpr env (In (App Ut.ArrLength _ [arr])) = do a' <- compileExpr env arr return $ arrayLength a' -- MutableReference compileExpr env (In (App Ut.GetRef _ [r])) = compileExpr env r -- NoInline -- Num -- Ord -- SizeProp compileExpr env (In (App Ut.PropSize _ [e])) = compileExpr env e -- SourceInfo compileExpr env (In (App (Ut.SourceInfo info) _ [a])) = do tellProg [Comment True info] compileExpr env a -- Tuple compileExpr env (In (App p _ [tup])) | p `elem` [ Ut.Sel1, Ut.Sel2, Ut.Sel3, Ut.Sel4, Ut.Sel5, Ut.Sel6, Ut.Sel7 , Ut.Sel8, Ut.Sel9, Ut.Sel10, Ut.Sel11, Ut.Sel12, Ut.Sel13 , Ut.Sel14, Ut.Sel15] = do tupExpr <- compileExpr env tup return $ StructField tupExpr ("member" ++ drop 3 (show p)) compileExpr env e@(In (App p _ _)) | p `elem` [ Ut.Parallel, Ut.SetLength, Ut.Sequential, Ut.Condition, Ut.ConditionM , Ut.MkFuture, Ut.Await, Ut.Bind, Ut.Then, Ut.Return, Ut.While, Ut.For, Ut.SetArr, Ut.EMaterialize , Ut.WhileLoop, Ut.ForLoop, Ut.RunMutableArray, Ut.NoInline , Ut.Switch, Ut.WithArray, Ut.Tup2, Ut.Tup3, Ut.Tup4, Ut.Tup5 , Ut.Tup6, Ut.Tup7, Ut.Tup8, Ut.Tup9, Ut.Tup10, Ut.Tup11, Ut.Tup11 , Ut.Tup12, Ut.Tup13, Ut.Tup14, Ut.Tup15] = compileProgFresh env e compileExpr env (In (App p t es)) = do es' <- mapM (compileExpr env) es return $ fun' (compileTypeRep (opts env) t) (compileOp p) es' compileExpr env e = compileProgFresh env e compileLet :: CompileEnv -> Ut.UntypedFeld -> Ut.Type -> Integer -> CodeWriter (Expression ()) compileLet env a ta v = do let var = mkVar (compileTypeRep (opts env) ta) v declare var compileProg env (Just var) a return var compileAssert :: CompileEnv -> Ut.UntypedFeld -> String -> CodeWriter () compileAssert env cond msg = do condExpr <- compileExpr env cond tellProg [call "assert" [ValueParameter condExpr]] unless (null msg) $ tellProg [Comment False $ "{" ++ msg ++ "}"] literal :: Options -> Ut.Lit -> CodeWriter (Expression ()) literal opt t@LUnit = return (ConstExpr $ literalConst opt t) literal opt t@LBool{} = return (ConstExpr $ literalConst opt t) literal opt t@LInt{} = return (ConstExpr $ literalConst opt t) literal opt t@LFloat{} = return (ConstExpr $ literalConst opt t) literal opt t@LDouble{} = return (ConstExpr $ literalConst opt t) literal opt t@LComplex{} = return (ConstExpr $ literalConst opt t) literal opt t@LArray{} = return (ConstExpr $ literalConst opt t) literal opt t = do loc <- freshVar opt "x" (typeof t) literalLoc opt loc t return loc -- | Returns true if we can represent the literal in Program. representableType :: Ut.Lit -> Bool representableType l | Ut.ArrayType{} <- t = True -- Simple types. | Ut.IntType{} <- t = True | Ut.ComplexType{} <- t = True | otherwise = t `elem` [Ut.UnitType, Ut.DoubleType, Ut.FloatType, Ut.BoolType] where t = typeof l literalConst :: Options -> Ut.Lit -> Constant () literalConst _ LUnit = IntConst 0 (Rep.NumType Ut.Unsigned Ut.S32) literalConst _ (LBool a) = BoolConst a literalConst _ (LInt s sz a) = IntConst (toInteger a) (Rep.NumType s sz) literalConst _ (LFloat a) = FloatConst a literalConst _ (LDouble a) = DoubleConst a literalConst opt (LArray t es) = ArrayConst (map (literalConst opt) es) $ compileTypeRep opt t literalConst opt (LComplex r i) = ComplexConst (literalConst opt r) (literalConst opt i) literalLoc :: Options -> Expression () -> Ut.Lit -> CodeWriter () literalLoc opt loc arr@Ut.LArray{} = tellProg [copyProg (Just loc) [ConstExpr $ literalConst opt arr]] literalLoc env loc (Ut.LTup2 ta tb) = do literalLoc env (StructField loc "member1") ta literalLoc env (StructField loc "member2") tb literalLoc env loc (Ut.LTup3 ta tb tc) = do literalLoc env (StructField loc "member1") ta literalLoc env (StructField loc "member2") tb literalLoc env (StructField loc "member3") tc literalLoc env loc (Ut.LTup4 ta tb tc td) = do literalLoc env (StructField loc "member1") ta literalLoc env (StructField loc "member2") tb literalLoc env (StructField loc "member3") tc literalLoc env (StructField loc "member4") td literalLoc env loc (Ut.LTup5 ta tb tc td te) = do literalLoc env (StructField loc "member1") ta literalLoc env (StructField loc "member2") tb literalLoc env (StructField loc "member3") tc literalLoc env (StructField loc "member4") td literalLoc env (StructField loc "member5") te literalLoc env loc (Ut.LTup6 ta tb tc td te tf) = do literalLoc env (StructField loc "member1") ta literalLoc env (StructField loc "member2") tb literalLoc env (StructField loc "member3") tc literalLoc env (StructField loc "member4") td literalLoc env (StructField loc "member5") te literalLoc env (StructField loc "member6") tf literalLoc env loc (Ut.LTup7 ta tb tc td te tf tg) = do literalLoc env (StructField loc "member1") ta literalLoc env (StructField loc "member2") tb literalLoc env (StructField loc "member3") tc literalLoc env (StructField loc "member4") td literalLoc env (StructField loc "member5") te literalLoc env (StructField loc "member6") tf literalLoc env (StructField loc "member7") tg literalLoc env loc (Ut.LTup8 ta tb tc td te tf tg th) = do literalLoc env (StructField loc "member1") ta literalLoc env (StructField loc "member2") tb literalLoc env (StructField loc "member3") tc literalLoc env (StructField loc "member4") td literalLoc env (StructField loc "member5") te literalLoc env (StructField loc "member6") tf literalLoc env (StructField loc "member7") tg literalLoc env (StructField loc "member8") th literalLoc env loc (Ut.LTup9 ta tb tc td te tf tg th ti) = do literalLoc env (StructField loc "member1") ta literalLoc env (StructField loc "member2") tb literalLoc env (StructField loc "member3") tc literalLoc env (StructField loc "member4") td literalLoc env (StructField loc "member5") te literalLoc env (StructField loc "member6") tf literalLoc env (StructField loc "member7") tg literalLoc env (StructField loc "member8") th literalLoc env (StructField loc "member9") ti literalLoc env loc (Ut.LTup10 ta tb tc td te tf tg th ti tj) = do literalLoc env (StructField loc "member1") ta literalLoc env (StructField loc "member2") tb literalLoc env (StructField loc "member3") tc literalLoc env (StructField loc "member4") td literalLoc env (StructField loc "member5") te literalLoc env (StructField loc "member6") tf literalLoc env (StructField loc "member7") tg literalLoc env (StructField loc "member8") th literalLoc env (StructField loc "member9") ti literalLoc env (StructField loc "member10") tj literalLoc env loc (Ut.LTup11 ta tb tc td te tf tg th ti tj tk) = do literalLoc env (StructField loc "member1") ta literalLoc env (StructField loc "member2") tb literalLoc env (StructField loc "member3") tc literalLoc env (StructField loc "member4") td literalLoc env (StructField loc "member5") te literalLoc env (StructField loc "member6") tf literalLoc env (StructField loc "member7") tg literalLoc env (StructField loc "member8") th literalLoc env (StructField loc "member9") ti literalLoc env (StructField loc "member10") tj literalLoc env (StructField loc "member11") tk literalLoc env loc (Ut.LTup12 ta tb tc td te tf tg th ti tj tk tl) = do literalLoc env (StructField loc "member1") ta literalLoc env (StructField loc "member2") tb literalLoc env (StructField loc "member3") tc literalLoc env (StructField loc "member4") td literalLoc env (StructField loc "member5") te literalLoc env (StructField loc "member6") tf literalLoc env (StructField loc "member7") tg literalLoc env (StructField loc "member8") th literalLoc env (StructField loc "member9") ti literalLoc env (StructField loc "member10") tj literalLoc env (StructField loc "member11") tk literalLoc env (StructField loc "member12") tl literalLoc env loc (Ut.LTup13 ta tb tc td te tf tg th ti tj tk tl tm) = do literalLoc env (StructField loc "member1") ta literalLoc env (StructField loc "member2") tb literalLoc env (StructField loc "member3") tc literalLoc env (StructField loc "member4") td literalLoc env (StructField loc "member5") te literalLoc env (StructField loc "member6") tf literalLoc env (StructField loc "member7") tg literalLoc env (StructField loc "member8") th literalLoc env (StructField loc "member9") ti literalLoc env (StructField loc "member10") tj literalLoc env (StructField loc "member11") tk literalLoc env (StructField loc "member12") tl literalLoc env (StructField loc "member13") tm literalLoc env loc (Ut.LTup14 ta tb tc td te tf tg th ti tj tk tl tm tn) = do literalLoc env (StructField loc "member1") ta literalLoc env (StructField loc "member2") tb literalLoc env (StructField loc "member3") tc literalLoc env (StructField loc "member4") td literalLoc env (StructField loc "member5") te literalLoc env (StructField loc "member6") tf literalLoc env (StructField loc "member7") tg literalLoc env (StructField loc "member8") th literalLoc env (StructField loc "member9") ti literalLoc env (StructField loc "member10") tj literalLoc env (StructField loc "member11") tk literalLoc env (StructField loc "member12") tl literalLoc env (StructField loc "member13") tm literalLoc env (StructField loc "member14") tn literalLoc env loc (Ut.LTup15 ta tb tc td te tf tg th ti tj tk tl tm tn to) = do literalLoc env (StructField loc "member1") ta literalLoc env (StructField loc "member2") tb literalLoc env (StructField loc "member3") tc literalLoc env (StructField loc "member4") td literalLoc env (StructField loc "member5") te literalLoc env (StructField loc "member6") tf literalLoc env (StructField loc "member7") tg literalLoc env (StructField loc "member8") th literalLoc env (StructField loc "member9") ti literalLoc env (StructField loc "member10") tj literalLoc env (StructField loc "member11") tk literalLoc env (StructField loc "member12") tl literalLoc env (StructField loc "member13") tm literalLoc env (StructField loc "member14") tn literalLoc env (StructField loc "member15") to literalLoc env loc t = do rhs <- literal env t assign (Just loc) rhs chaseTree :: CompileEnv -> Location -> Ut.UntypedFeld -> Ut.UntypedFeld -> CodeWriter [(Pattern (), Block ())] chaseTree env loc _s (In (App Ut.Condition _ [In (App Ut.Equal _ [c, _]), t, f])) -- , alphaEq s a -- TODO check that the scrutinees are equal = do e <- compileExpr env c (_,body) <- confiscateBlock $ compileProg env loc t cases <- chaseTree env loc _s f return $ (Pat e, body) : cases chaseTree env loc _ a = do (_,body) <- confiscateBlock $ compileProg env loc a return [(PatDefault, body)] -- | Chase down the right-spine of `Bind` and `Then` constructs and return -- the last term chaseBind :: Ut.UntypedFeld -> Ut.UntypedFeld chaseBind (In (App Ut.Bind _ [_, In (Ut.Lambda _ body)])) = chaseBind body chaseBind (In (App Ut.Then _ [_, body])) = chaseBind body chaseBind a = a {- NOTES: The trick of doing a copy at the end, i.e. `tellProg [copyProg loc e]`, when compiling WithArray is important. It allows us to safely return the pure array that is passed in as input. This is nice because it allows us to implement `freezeArray` in terms of `withArray`. In most cases I expect `withArray` to return a scalar as its final result and then the copyProg is harmless. -} compileBind :: CompileEnv -> (Ut.Var, Ut.UntypedFeld) -> CodeWriter () compileBind env (Ut.Var v t, e) = do let var = mkVar (compileTypeRep (opts env) t) v declare var compileProg env (Just var) e -- | Translates Op names to strings. compileOp :: Ut.Op -> String -- Bits compileOp Ut.BAnd = "&" compileOp Ut.BOr = "|" compileOp Ut.BXor = "^" -- Complex compileOp Ut.RealPart = "creal" compileOp Ut.ImagPart = "cimag" compileOp Ut.Sign = "signum" -- Eq compileOp Ut.Equal = "==" compileOp Ut.NotEqual = "/=" -- FFI compileOp (Ut.ForeignImport s) = s -- Floating compileOp Ut.Exp = "exp" -- Fractional compileOp Ut.DivFrac = "/" -- Integral compileOp Ut.IExp = "pow" -- Logic compileOp Ut.And = "&&" compileOp Ut.Or = "||" -- Num compileOp Ut.Add = "+" compileOp Ut.Sub = "-" compileOp Ut.Mul = "*" -- Ord compileOp Ut.LTH = "<" compileOp Ut.GTH = ">" compileOp Ut.LTE = "<=" compileOp Ut.GTE = ">=" compileOp p = toLower h:t where (h:t) = show p
emwap/feldspar-compiler
lib/Feldspar/Compiler/Imperative/FromCore.hs
bsd-3-clause
50,193
0
21
12,032
19,634
9,572
10,062
-1
-1
module Main where import System.Exit import System.Environment import System.Console.GetOpt import Control.Monad (forM_, when, unless) import Control.Monad.Trans import Control.Applicative ((<$>)) import Tools.Quarry import Data.Word import Data.List import Data.Maybe import System.IO import System.Process data SubCommand = Init | Import | Set | Get | Find | Tags | Cats | Info | Exist | View deriving (Show,Eq) data InitOpt = InitHelp deriving (Show,Eq) data ImportOpt = ImportHelp | ImportTy ImportType | ImportDate String | ImportTag String | ImportFilename String deriving (Show,Eq) data FindOpt = FindHelp deriving (Show,Eq) data TagOpt = TagHelp | TagCategory String deriving (Show,Eq) data CatsOpt = CatsHelp deriving (Show,Eq) data ExistOpt = ExistHelp | ExistMissing | ExistAlready deriving (Show,Eq) data ViewOpt = ViewHelp deriving (Show,Eq) usage Init = error "usage: quarry init <repository-path>" usage Import = error "usage: quarry import <repository-path> <file>" usage Set = error "usage: quarry set <repository-path> <digest> [+/-tag]" usage Get = error "usage: quarry get <repository-path> <digest>" usage Find = error "usage: quarry find <repository-path> <query>" usage Tags = error "usage: quarry tags [--category <category>] <repository-path> <tag prefix> ..." usage Cats = error "usage: quarry cats <add|list> <repository-path> <cat prefix> ..." usage Exist = error "usage: quarry exists <repository-path> <file> ..." usage Info = error "usage: quarry info <repository-path>" usage View = error "usage: quarry view <repository-path> <digest>" reportOptError errOpts | null errOpts = return () | otherwise = mapM_ (putStrLn . ("parseError: " ++)) errOpts >> exitFailure fromTagArg :: String -> Either TagName Tag fromTagArg s = case break (== ':') s of (_,"") -> Left s (_ ,[':']) -> error "empty tag, expecting 'category:tag', got 'category:'" (cat,':':t) -> Right $ Tag { tagName = t, tagCat = cat } _ -> error "impossible with break" tag s = maybe (error $ "cannot resolve " ++ s) id <$> resolveTag (fromTagArg s) -- create a tag object, automatically filling category to personal if cannot be found. resolveAddTag s = do let t = fromTagArg s r <- resolveTag t case r of Just _ -> return r Nothing -> case t of Left _ -> return $ Just $ Tag { tagName = s, tagCat = "personal" } _ -> return r cmdInit args = do let (optArgs, nonOpts, errOpts) = getOpt Permute options args when (InitHelp `elem` optArgs) $ do usage Init >> exitSuccess reportOptError errOpts case nonOpts of [path] -> initialize True path >> return () _ -> usage Init where options = [ Option ['h'] ["help"] (NoArg InitHelp) "show help" ] -- | import a file cmdImport args = do let (optArgs, nonOpts, errOpts) = getOpt Permute options args when (ImportHelp `elem` optArgs) $ do usage Import >> exitSuccess reportOptError errOpts case nonOpts of [path,file] -> doImport optArgs path file _ -> usage Import where options = [ Option ['h'] ["help"] (NoArg ImportHelp) "show help" , Option ['s'] ["symlink"] (NoArg (ImportTy ImportSymlink)) "use a symlink to import into the hashfs" , Option [] ["hardlink"] (NoArg (ImportTy ImportHardlink)) "use a hardlink to import into the hashfs" , Option ['d'] ["date"] (ReqArg ImportDate "date") "add a date in posix seconds" , Option ['t'] ["tag"] (ReqArg ImportTag "tag") "add a tag" , Option ['f'] ["filename"] (ReqArg ImportFilename "filename") "override the filename" ] hardcodedDataCat = CategoryPersonal doImport optArgs path file = do let (date,tags,mFilename,ity) = foldl (\acc@(d,accTags,accFile,t) f -> case f of ImportTy ty -> (d,accTags,accFile,ty) ImportDate da -> (read da,accTags,accFile,t) ImportTag ta -> (d,ta:accTags,accFile,t) ImportFilename fi -> (d,accTags,Just fi,t) _ -> acc) (0 :: Word64, [], Nothing, ImportCopy) optArgs let mDate = case date of 0 -> Nothing _ -> Just $ fromIntegral date conf <- initialize False path (digest,isNew) <- runQuarry conf $ do addTags <- catMaybes <$> mapM resolveAddTag tags importFile ity hardcodedDataCat mDate (("/old/" ++) `fmap` mFilename) addTags file if isNew then putStrLn (show digest) else hPutStrLn stderr (show digest ++ " already existing ") cmdSet args = case args of path:digest:tagPatches -> doSet path (maybe (error "not a valid digest") id $ readDigest digest) tagPatches _ -> usage Import where doSet path digest tagPatches = do let (addTagArgs, delTagArgs) = foldl readPatch ([], []) tagPatches conf <- initialize False path runQuarry conf $ do --tag s = maybe (error $ "cannot resolve " ++ s) id <$> resolveTag (fromTagArg s) addTags <- catMaybes <$> mapM resolveAddTag addTagArgs delTags <- catMaybes <$> mapM (resolveTag . fromTagArg) delTagArgs liftIO $ putStrLn ("deleting tags: " ++ show delTags) liftIO $ putStrLn ("adding tags: " ++ show addTags) updateDigest digest addTags delTags readPatch (a,d) s = case s of '+':toAdd -> (toAdd:a, d) '-':toRem -> (a, toRem:d) _ -> (a,d) cmdGet _ = error ("get not implemented") cmdFind args = do let (optArgs, nonOpts, errOpts) = getOpt Permute options args when (FindHelp `elem` optArgs) $ do usage Find >> exitSuccess reportOptError errOpts case nonOpts of path:tags -> doFind path tags _ -> usage Find where options = [ Option ['h'] ["help"] (NoArg FindHelp) "show help" ] doFind path tagArgs = do conf <- initialize False path runQuarry conf $ do tags <- mapM tag tagArgs digests <- findDigestWithTags tags liftIO $ mapM_ (putStrLn . show) digests cmdTags args = do let (optArgs, nonOpts, errOpts) = getOpt Permute options args when (TagHelp `elem` optArgs) $ do usage Tags >> exitSuccess reportOptError errOpts case nonOpts of [] -> usage Tags path:l -> doTags optArgs path l where options = [ Option ['h'] ["help"] (NoArg TagHelp) "show help" , Option ['c'] ["category"] (ReqArg TagCategory "category") "restrict tag search to a specific category" ] doTags optArgs path l = do let cat = foldl (\acc f -> case f of TagCategory c -> Just c _ -> acc) Nothing optArgs conf <- initialize False path runQuarry conf $ forM_ l $ \s -> do tags <- findTags (Just s) cat mapM_ (liftIO . putStrLn . show) tags cmdExist args = do let (optArgs, nonOpts, errOpts) = getOpt Permute options args when (ExistHelp `elem` optArgs) $ do usage Exist >> exitSuccess when (ExistMissing `elem` optArgs && ExistAlready `elem` optArgs) $ putStrLn "cannot specify missing and exist at the same time" >> exitFailure reportOptError errOpts case nonOpts of path:f1:fs -> doExist optArgs path (f1:fs) _ -> usage Exist where options = [ Option ['h'] ["help"] (NoArg ExistHelp) "show help" , Option ['m'] ["missing"] (NoArg ExistMissing) "show only missing file from the store" , Option ['e'] ["exist"] (NoArg ExistAlready) "show only already existing file in the store" ] doExist optArgs path filenames = do conf <- initialize False path runQuarry conf $ mapM_ (check optArgs) filenames check optArgs filename = do d <- computeDigest filename x <- exist d liftIO $ do case (ExistAlready `elem` optArgs, ExistMissing `elem` optArgs) of (True, True) -> error "impossible" (False, True) -> when x $ putStrLn filename (True, False) -> unless x $ putStrLn filename (False, False) -> if x then putStrLn (filename ++ " " ++ show d) else putStrLn (filename ++ " MISSING") cmdCats args = do let (optArgs, nonOpts, errOpts) = getOpt Permute options args when (CatsHelp `elem` optArgs) $ do usage Cats >> exitSuccess reportOptError errOpts case nonOpts of [] -> usage Cats "add":path:name:[] -> doCatAdd optArgs path name "list":path:_ -> doCats optArgs path _:_ -> error ("error unknown sub command " ++ show nonOpts ++ " in cats: expected list or add") where options = [ Option ['h'] ["help"] (NoArg CatsHelp) "show help" ] doCatAdd _ path name = do conf <- initialize False path runQuarry conf $ addCategory name doCats _ path = do conf <- initialize False path runQuarry conf $ do cats <- getCategoryTable mapM_ (\(_,(c,a)) -> liftIO $ putStrLn (c ++ " (abstract=" ++ show a ++ ")")) cats cmdInfo args = do case args of path:[] -> doInfo path _ -> usage Import where doInfo path = do conf <- initialize False path info <- runQuarry conf $ getInfo putStrLn ("files : " ++ show (infoNFile info)) putStrLn ("tags : " ++ show (infoNTag info)) putStrLn ("categories : " ++ show (infoNCategory info)) cmdView args = do case args of path:digests -> doView path digests _ -> usage View where doView path digests_ = do let digests = catMaybes $ map readDigest digests_ conf <- initialize False path l <- runQuarry conf $ mapM getPathAndType $ digests forM_ l $ \(p, ty) -> do case ty of QuarryTypeImage -> do ec <- rawSystem "eog" [p] unless (ec == ExitSuccess) $ exitWith ec _ -> putStrLn (show ty ++ " " ++ show p) getPathAndType digest = do path <- getDigestPath digest ty <- getQuarryFileType path return (path, ty) commands = [ ("init" , (cmdInit, "initialize a repository")) , ("import", (cmdImport, "import a file into quarry")) , ("set" , (cmdSet, "set some metadata")) , ("tags" , (cmdTags, "list tags")) , ("cats" , (cmdCats, "list categories")) , ("get" , (cmdGet, "get some metadata")) , ("find" , (cmdFind, "find contents by query")) , ("exist" , (cmdExist, "check if some file already exist")) , ("info" , (cmdInfo, "get quarry general state information")) , ("view" , (cmdView, "use a viewer on digest")) ] main = do args <- getArgs case args of [] -> error ("expecting one subcommand of: " ++ intercalate ", " (map fst commands)) cmd:subArgs -> case lookup cmd commands of Nothing -> error ("unrecognized command: " ++ cmd) Just (fCommand, _) -> fCommand subArgs
vincenthz/quarry
src/Quarry.hs
bsd-3-clause
12,085
0
21
4,134
3,644
1,847
1,797
246
8
{-# LANGUAGE GeneralizedNewtypeDeriving #-} module Jana.Types ( Array, Stack, Value(..), nil, performOperation, performModOperation, showValueType, typesMatch, truthy, Store, printVdecl, showStore, emptyStore, storeFromList, getRef, getVar, getRefValue, bindVar, unbindVar, setVar, EvalEnv(..), EvalOptions(..), defaultOptions, ProcEnv, emptyProcEnv, procEnvFromList, getProc, Eval, runEval, (<!!>) ) where import Prelude hiding (GT, LT, EQ) import Data.Bits import Data.List (intercalate) import Data.IORef import Control.Monad.State import Control.Monad.Reader import Control.Monad.Except import Text.Printf (printf) import qualified Data.Maybe as Maybe import qualified Data.Map as Map import Text.Parsec.Pos import Jana.Aliases import Jana.Ast import Jana.Error import Jana.ErrorMessages type Array = [Integer] type Stack = [Integer] -- Types of values an expression can evaluate to. data Value = JInt Integer | JBool Bool | JArray Array | JStack Stack deriving (Eq) instance Show Value where show (JInt x) = show x show (JArray xs) = "{" ++ intercalate ", " (map show xs) ++ "}" show (JStack []) = "nil" show (JStack xs) = "<" ++ intercalate ", " (map show xs) ++ "]" showValueType :: Value -> String showValueType (JInt _) = "int" showValueType (JStack _) = "stack" showValueType (JArray _) = "array" showValueType (JBool _) = "bool" typesMatch :: Value -> Value -> Bool typesMatch (JInt _) (JInt _) = True typesMatch (JArray _) (JArray _) = True typesMatch (JStack _) (JStack _) = True typesMatch (JBool _) (JBool _) = True typesMatch _ _ = False nil = JStack [] truthy :: Value -> Bool truthy (JInt 0) = False truthy (JStack []) = False truthy _ = True boolToInt :: Num a => (a -> a -> Bool) -> a -> a -> a boolToInt f x y = if f x y then 1 else 0 wrap :: (a -> Value) -> (Integer -> Integer -> a) -> Integer -> Integer -> Value wrap m f x y = m $ f x y opFunc :: BinOp -> Integer -> Integer -> Value opFunc Add = wrap JInt (+) opFunc Sub = wrap JInt (-) opFunc Mul = wrap JInt (*) opFunc Div = wrap JInt div opFunc Mod = wrap JInt mod opFunc And = wrap JInt (.&.) opFunc Or = wrap JInt (.|.) opFunc Xor = wrap JInt xor opFunc LAnd = undefined -- handled by evalExpr opFunc LOr = undefined -- handled by evalExpr opFunc GT = wrap JBool (>) opFunc LT = wrap JBool (<) opFunc EQ = wrap JBool (==) opFunc NEQ = wrap JBool (/=) opFunc GE = wrap JBool (>=) opFunc LE = wrap JBool (<=) performOperation :: BinOp -> Value -> Value -> SourcePos -> SourcePos -> Eval Value performOperation Div (JInt _) (JInt 0) _ pos = pos <!!> divisionByZero performOperation op (JInt x) (JInt y) _ _ = return $ opFunc op x y performOperation _ (JInt _) val _ pos = pos <!!> typeMismatch ["int"] (showValueType val) performOperation _ val _ pos _ = pos <!!> typeMismatch ["int"] (showValueType val) performModOperation :: ModOp -> Value -> Value -> SourcePos -> SourcePos -> Eval Value performModOperation modOp = performOperation $ modOpToBinOp modOp where modOpToBinOp AddEq = Add modOpToBinOp SubEq = Sub modOpToBinOp XorEq = Xor -- -- Environment -- type Store = Map.Map String (IORef Value) printVdecl :: String -> Value -> String printVdecl name val@(JArray xs) = printf "%s[%d] = %s" name (length xs) (show val) printVdecl name val = printf "%s = %s" name (show val) showStore :: Store -> IO String showStore store = liftM (intercalate "\n") (mapM (\(name, ref) -> liftM (printVdecl name) (readIORef ref)) (Map.toList store)) emptyStore = Map.empty storeFromList :: [(String, IORef Value)] -> Store storeFromList = Map.fromList getRef :: Ident -> Eval (IORef Value) getRef (Ident name pos) = do storeEnv <- get case Map.lookup name storeEnv of Just ref -> return ref Nothing -> pos <!!> unboundVar name getVar :: Ident -> Eval Value getVar id = getRef id >>= liftIO . readIORef getRefValue :: IORef Value -> Eval Value getRefValue = liftIO . readIORef -- Bind a variable name to a new reference bindVar :: Ident -> Value -> Eval () bindVar (Ident name pos) val = do storeEnv <- get ref <- liftIO $ newIORef val case Map.lookup name storeEnv of Nothing -> put $ Map.insert name ref storeEnv Just _ -> pos <!!> alreadyBound name unbindVar :: Ident -> Eval () unbindVar = modify . Map.delete . ident -- Set the value of a variable (modifying the reference) setVar :: Ident -> Value -> Eval () setVar id val = do ref <- getRef id liftIO $ writeIORef ref val data EvalEnv = EE { evalOptions :: EvalOptions , procEnv :: ProcEnv , aliases :: AliasSet } data EvalOptions = EvalOptions { modInt :: Bool, runReverse :: Bool } defaultOptions = EvalOptions { modInt = False, runReverse = False } type ProcEnv = Map.Map String Proc emptyProcEnv = Map.empty procEnvFromList :: [Proc] -> Either JanaError ProcEnv procEnvFromList = foldM insertProc emptyProcEnv where insertProc env p = if Map.notMember (ident p) env then if checkDuplicateArgs (makeIdentList p) then Right (Map.insert (ident p) p env) else Left $ newErrorMessage (ppos p) (procDuplicateArgs p) else Left $ newErrorMessage (ppos p) (procDefined p) ppos Proc { procname = (Ident _ pos) } = pos makeIdentList :: Proc -> [Ident] makeIdentList (Proc {params = params}) = map getVdeclIdent params where getVdeclIdent (Scalar _ id _) = id getVdeclIdent (Array id _ _) = id checkDuplicateArgs :: [Ident] -> Bool checkDuplicateArgs [] = True checkDuplicateArgs ([arg]) = True checkDuplicateArgs (arg:args) = arg `notElem` args && checkDuplicateArgs args getProc :: Ident -> Eval Proc getProc (Ident funName pos) = do when (funName == "main") $ pos <!!> callingMainError procEnv <- asks procEnv case Map.lookup funName procEnv of Just proc -> return proc Nothing -> pos <!!> undefProc funName -- -- Evaluation -- instance Functor Eval where fmap = liftM instance Applicative Eval where pure = return (<*>) = ap -- defined in Control.Monad newtype Eval a = E { runE :: StateT Store (ReaderT EvalEnv (ExceptT JanaError IO)) a } deriving (Monad, MonadIO, MonadError JanaError, MonadReader EvalEnv, MonadState Store) runEval :: Eval a -> Store -> EvalEnv -> IO (Either JanaError (a, Store)) runEval eval store procs = runExceptT (runReaderT (runStateT (runE eval) store) procs) throwJanaError :: SourcePos -> Message -> Eval a throwJanaError pos msg = throwError $ newErrorMessage pos msg infixr 1 <!!> pos <!!> msg = throwJanaError pos msg
tyoko-dev/jana
src/Jana/Types.hs
bsd-3-clause
6,785
0
13
1,569
2,470
1,298
1,172
169
3
{- % (c) The University of Glasgow 2006 (c) The GRASP/AQUA Project, Glasgow University, 1992-1998 TcGenDeriv: Generating derived instance declarations This module is nominally ``subordinate'' to @TcDeriv@, which is the ``official'' interface to deriving-related things. This is where we do all the grimy bindings' generation. -} {-# LANGUAGE CPP, ScopedTypeVariables #-} {-# LANGUAGE FlexibleContexts #-} module TcGenDeriv ( BagDerivStuff, DerivStuff(..), hasBuiltinDeriving, FFoldType(..), functorLikeTraverse, deepSubtypesContaining, foldDataConArgs, mkCoerceClassMethEqn, gen_Newtype_binds, genAuxBinds, ordOpTbl, boxConTbl, litConTbl, mkRdrFunBind ) where #include "HsVersions.h" import LoadIface( loadInterfaceForName ) import HscTypes( lookupFixity, mi_fix ) import TcRnMonad import HsSyn import RdrName import BasicTypes import Module( getModule ) import DataCon import Name import Fingerprint import Encoding import DynFlags import PrelInfo import FamInstEnv( FamInst ) import PrelNames import THNames import Module ( moduleName, moduleNameString , moduleUnitId, unitIdString ) import MkId ( coerceId ) import PrimOp import SrcLoc import TyCon import TcType import TysPrim import TysWiredIn import Type import Class import TyCoRep import VarSet import VarEnv import State import Util import Var import Outputable import Lexeme import FastString import Pair import Bag import TcEnv (InstInfo) import StaticFlags( opt_PprStyle_Debug ) import ListSetOps ( assocMaybe ) import Data.List ( partition, intersperse ) import Data.Maybe ( catMaybes, isJust ) type BagDerivStuff = Bag DerivStuff data AuxBindSpec = DerivCon2Tag TyCon -- The con2Tag for given TyCon | DerivTag2Con TyCon -- ...ditto tag2Con | DerivMaxTag TyCon -- ...and maxTag deriving( Eq ) -- All these generate ZERO-BASED tag operations -- I.e first constructor has tag 0 data DerivStuff -- Please add this auxiliary stuff = DerivAuxBind AuxBindSpec -- Generics | DerivFamInst FamInst -- New type family instances -- New top-level auxiliary bindings | DerivHsBind (LHsBind RdrName, LSig RdrName) -- Also used for SYB | DerivInst (InstInfo RdrName) -- New, auxiliary instances {- ************************************************************************ * * Class deriving diagnostics * * ************************************************************************ Only certain blessed classes can be used in a deriving clause. These classes are listed below in the definition of hasBuiltinDeriving (with the exception of Generic and Generic1, which are handled separately in TcGenGenerics). A class might be able to be used in a deriving clause if it -XDeriveAnyClass is willing to support it. The canDeriveAnyClass function checks if this is the case. -} hasBuiltinDeriving :: Class -> Maybe (SrcSpan -> TyCon -> TcM (LHsBinds RdrName, BagDerivStuff)) hasBuiltinDeriving clas = assocMaybe gen_list (getUnique clas) where gen_list :: [(Unique, SrcSpan -> TyCon -> TcM (LHsBinds RdrName, BagDerivStuff))] gen_list = [ (eqClassKey, simple gen_Eq_binds) , (ordClassKey, simple gen_Ord_binds) , (enumClassKey, simple gen_Enum_binds) , (boundedClassKey, simple gen_Bounded_binds) , (ixClassKey, simple gen_Ix_binds) , (showClassKey, with_fix_env gen_Show_binds) , (readClassKey, with_fix_env gen_Read_binds) , (dataClassKey, gen_Data_binds) , (functorClassKey, simple gen_Functor_binds) , (foldableClassKey, simple gen_Foldable_binds) , (traversableClassKey, simple gen_Traversable_binds) , (liftClassKey, simple gen_Lift_binds) ] simple gen_fn loc tc = return (gen_fn loc tc) with_fix_env gen_fn loc tc = do { fix_env <- getDataConFixityFun tc ; return (gen_fn fix_env loc tc) } getDataConFixityFun :: TyCon -> TcM (Name -> Fixity) -- If the TyCon is locally defined, we want the local fixity env; -- but if it is imported (which happens for standalone deriving) -- we need to get the fixity env from the interface file -- c.f. RnEnv.lookupFixity, and Trac #9830 getDataConFixityFun tc = do { this_mod <- getModule ; if nameIsLocalOrFrom this_mod name then do { fix_env <- getFixityEnv ; return (lookupFixity fix_env) } else do { iface <- loadInterfaceForName doc name -- Should already be loaded! ; return (mi_fix iface . nameOccName) } } where name = tyConName tc doc = text "Data con fixities for" <+> ppr name {- ************************************************************************ * * Eq instances * * ************************************************************************ Here are the heuristics for the code we generate for @Eq@. Let's assume we have a data type with some (possibly zero) nullary data constructors and some ordinary, non-nullary ones (the rest, also possibly zero of them). Here's an example, with both \tr{N}ullary and \tr{O}rdinary data cons. data Foo ... = N1 | N2 ... | Nn | O1 a b | O2 Int | O3 Double b b | ... * For the ordinary constructors (if any), we emit clauses to do The Usual Thing, e.g.,: (==) (O1 a1 b1) (O1 a2 b2) = a1 == a2 && b1 == b2 (==) (O2 a1) (O2 a2) = a1 == a2 (==) (O3 a1 b1 c1) (O3 a2 b2 c2) = a1 == a2 && b1 == b2 && c1 == c2 Note: if we're comparing unlifted things, e.g., if 'a1' and 'a2' are Float#s, then we have to generate case (a1 `eqFloat#` a2) of r -> r for that particular test. * If there are a lot of (more than en) nullary constructors, we emit a catch-all clause of the form: (==) a b = case (con2tag_Foo a) of { a# -> case (con2tag_Foo b) of { b# -> case (a# ==# b#) of { r -> r }}} If con2tag gets inlined this leads to join point stuff, so it's better to use regular pattern matching if there aren't too many nullary constructors. "Ten" is arbitrary, of course * If there aren't any nullary constructors, we emit a simpler catch-all: (==) a b = False * For the @(/=)@ method, we normally just use the default method. If the type is an enumeration type, we could/may/should? generate special code that calls @con2tag_Foo@, much like for @(==)@ shown above. We thought about doing this: If we're also deriving 'Ord' for this tycon, we generate: instance ... Eq (Foo ...) where (==) a b = case (compare a b) of { _LT -> False; _EQ -> True ; _GT -> False} (/=) a b = case (compare a b) of { _LT -> True ; _EQ -> False; _GT -> True } However, that requires that (Ord <whatever>) was put in the context for the instance decl, which it probably wasn't, so the decls produced don't get through the typechecker. -} gen_Eq_binds :: SrcSpan -> TyCon -> (LHsBinds RdrName, BagDerivStuff) gen_Eq_binds loc tycon = (method_binds, aux_binds) where all_cons = tyConDataCons tycon (nullary_cons, non_nullary_cons) = partition isNullarySrcDataCon all_cons -- If there are ten or more (arbitrary number) nullary constructors, -- use the con2tag stuff. For small types it's better to use -- ordinary pattern matching. (tag_match_cons, pat_match_cons) | nullary_cons `lengthExceeds` 10 = (nullary_cons, non_nullary_cons) | otherwise = ([], all_cons) no_tag_match_cons = null tag_match_cons fall_through_eqn | no_tag_match_cons -- All constructors have arguments = case pat_match_cons of [] -> [] -- No constructors; no fall-though case [_] -> [] -- One constructor; no fall-though case _ -> -- Two or more constructors; add fall-through of -- (==) _ _ = False [([nlWildPat, nlWildPat], false_Expr)] | otherwise -- One or more tag_match cons; add fall-through of -- extract tags compare for equality = [([a_Pat, b_Pat], untag_Expr tycon [(a_RDR,ah_RDR), (b_RDR,bh_RDR)] (genPrimOpApp (nlHsVar ah_RDR) eqInt_RDR (nlHsVar bh_RDR)))] aux_binds | no_tag_match_cons = emptyBag | otherwise = unitBag $ DerivAuxBind $ DerivCon2Tag tycon method_binds = listToBag [eq_bind, ne_bind] eq_bind = mk_FunBind loc eq_RDR (map pats_etc pat_match_cons ++ fall_through_eqn) ne_bind = mk_easy_FunBind loc ne_RDR [a_Pat, b_Pat] ( nlHsApp (nlHsVar not_RDR) (nlHsPar (nlHsVarApps eq_RDR [a_RDR, b_RDR]))) ------------------------------------------------------------------ pats_etc data_con = let con1_pat = nlConVarPat data_con_RDR as_needed con2_pat = nlConVarPat data_con_RDR bs_needed data_con_RDR = getRdrName data_con con_arity = length tys_needed as_needed = take con_arity as_RDRs bs_needed = take con_arity bs_RDRs tys_needed = dataConOrigArgTys data_con in ([con1_pat, con2_pat], nested_eq_expr tys_needed as_needed bs_needed) where nested_eq_expr [] [] [] = true_Expr nested_eq_expr tys as bs = foldl1 and_Expr (zipWith3Equal "nested_eq" nested_eq tys as bs) where nested_eq ty a b = nlHsPar (eq_Expr tycon ty (nlHsVar a) (nlHsVar b)) {- ************************************************************************ * * Ord instances * * ************************************************************************ Note [Generating Ord instances] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Suppose constructors are K1..Kn, and some are nullary. The general form we generate is: * Do case on first argument case a of K1 ... -> rhs_1 K2 ... -> rhs_2 ... Kn ... -> rhs_n _ -> nullary_rhs * To make rhs_i If i = 1, 2, n-1, n, generate a single case. rhs_2 case b of K1 {} -> LT K2 ... -> ...eq_rhs(K2)... _ -> GT Otherwise do a tag compare against the bigger range (because this is the one most likely to succeed) rhs_3 case tag b of tb -> if 3 <# tg then GT else case b of K3 ... -> ...eq_rhs(K3).... _ -> LT * To make eq_rhs(K), which knows that a = K a1 .. av b = K b1 .. bv we just want to compare (a1,b1) then (a2,b2) etc. Take care on the last field to tail-call into comparing av,bv * To make nullary_rhs generate this case con2tag a of a# -> case con2tag b of -> a# `compare` b# Several special cases: * Two or fewer nullary constructors: don't generate nullary_rhs * Be careful about unlifted comparisons. When comparing unboxed values we can't call the overloaded functions. See function unliftedOrdOp Note [Do not rely on compare] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ It's a bad idea to define only 'compare', and build the other binary comparisons on top of it; see Trac #2130, #4019. Reason: we don't want to laboriously make a three-way comparison, only to extract a binary result, something like this: (>) (I# x) (I# y) = case <# x y of True -> False False -> case ==# x y of True -> False False -> True So for sufficiently small types (few constructors, or all nullary) we generate all methods; for large ones we just use 'compare'. -} data OrdOp = OrdCompare | OrdLT | OrdLE | OrdGE | OrdGT ------------ ordMethRdr :: OrdOp -> RdrName ordMethRdr op = case op of OrdCompare -> compare_RDR OrdLT -> lt_RDR OrdLE -> le_RDR OrdGE -> ge_RDR OrdGT -> gt_RDR ------------ ltResult :: OrdOp -> LHsExpr RdrName -- Knowing a<b, what is the result for a `op` b? ltResult OrdCompare = ltTag_Expr ltResult OrdLT = true_Expr ltResult OrdLE = true_Expr ltResult OrdGE = false_Expr ltResult OrdGT = false_Expr ------------ eqResult :: OrdOp -> LHsExpr RdrName -- Knowing a=b, what is the result for a `op` b? eqResult OrdCompare = eqTag_Expr eqResult OrdLT = false_Expr eqResult OrdLE = true_Expr eqResult OrdGE = true_Expr eqResult OrdGT = false_Expr ------------ gtResult :: OrdOp -> LHsExpr RdrName -- Knowing a>b, what is the result for a `op` b? gtResult OrdCompare = gtTag_Expr gtResult OrdLT = false_Expr gtResult OrdLE = false_Expr gtResult OrdGE = true_Expr gtResult OrdGT = true_Expr ------------ gen_Ord_binds :: SrcSpan -> TyCon -> (LHsBinds RdrName, BagDerivStuff) gen_Ord_binds loc tycon | null tycon_data_cons -- No data-cons => invoke bale-out case = (unitBag $ mk_FunBind loc compare_RDR [], emptyBag) | otherwise = (unitBag (mkOrdOp OrdCompare) `unionBags` other_ops, aux_binds) where aux_binds | single_con_type = emptyBag | otherwise = unitBag $ DerivAuxBind $ DerivCon2Tag tycon -- Note [Do not rely on compare] other_ops | (last_tag - first_tag) <= 2 -- 1-3 constructors || null non_nullary_cons -- Or it's an enumeration = listToBag (map mkOrdOp [OrdLT,OrdLE,OrdGE,OrdGT]) | otherwise = emptyBag get_tag con = dataConTag con - fIRST_TAG -- We want *zero-based* tags, because that's what -- con2Tag returns (generated by untag_Expr)! tycon_data_cons = tyConDataCons tycon single_con_type = isSingleton tycon_data_cons (first_con : _) = tycon_data_cons (last_con : _) = reverse tycon_data_cons first_tag = get_tag first_con last_tag = get_tag last_con (nullary_cons, non_nullary_cons) = partition isNullarySrcDataCon tycon_data_cons mkOrdOp :: OrdOp -> LHsBind RdrName -- Returns a binding op a b = ... compares a and b according to op .... mkOrdOp op = mk_easy_FunBind loc (ordMethRdr op) [a_Pat, b_Pat] (mkOrdOpRhs op) mkOrdOpRhs :: OrdOp -> LHsExpr RdrName mkOrdOpRhs op -- RHS for comparing 'a' and 'b' according to op | length nullary_cons <= 2 -- Two nullary or fewer, so use cases = nlHsCase (nlHsVar a_RDR) $ map (mkOrdOpAlt op) tycon_data_cons -- i.e. case a of { C1 x y -> case b of C1 x y -> ....compare x,y... -- C2 x -> case b of C2 x -> ....comopare x.... } | null non_nullary_cons -- All nullary, so go straight to comparing tags = mkTagCmp op | otherwise -- Mixed nullary and non-nullary = nlHsCase (nlHsVar a_RDR) $ (map (mkOrdOpAlt op) non_nullary_cons ++ [mkHsCaseAlt nlWildPat (mkTagCmp op)]) mkOrdOpAlt :: OrdOp -> DataCon -> LMatch RdrName (LHsExpr RdrName) -- Make the alternative (Ki a1 a2 .. av -> mkOrdOpAlt op data_con = mkHsCaseAlt (nlConVarPat data_con_RDR as_needed) (mkInnerRhs op data_con) where as_needed = take (dataConSourceArity data_con) as_RDRs data_con_RDR = getRdrName data_con mkInnerRhs op data_con | single_con_type = nlHsCase (nlHsVar b_RDR) [ mkInnerEqAlt op data_con ] | tag == first_tag = nlHsCase (nlHsVar b_RDR) [ mkInnerEqAlt op data_con , mkHsCaseAlt nlWildPat (ltResult op) ] | tag == last_tag = nlHsCase (nlHsVar b_RDR) [ mkInnerEqAlt op data_con , mkHsCaseAlt nlWildPat (gtResult op) ] | tag == first_tag + 1 = nlHsCase (nlHsVar b_RDR) [ mkHsCaseAlt (nlConWildPat first_con) (gtResult op) , mkInnerEqAlt op data_con , mkHsCaseAlt nlWildPat (ltResult op) ] | tag == last_tag - 1 = nlHsCase (nlHsVar b_RDR) [ mkHsCaseAlt (nlConWildPat last_con) (ltResult op) , mkInnerEqAlt op data_con , mkHsCaseAlt nlWildPat (gtResult op) ] | tag > last_tag `div` 2 -- lower range is larger = untag_Expr tycon [(b_RDR, bh_RDR)] $ nlHsIf (genPrimOpApp (nlHsVar bh_RDR) ltInt_RDR tag_lit) (gtResult op) $ -- Definitely GT nlHsCase (nlHsVar b_RDR) [ mkInnerEqAlt op data_con , mkHsCaseAlt nlWildPat (ltResult op) ] | otherwise -- upper range is larger = untag_Expr tycon [(b_RDR, bh_RDR)] $ nlHsIf (genPrimOpApp (nlHsVar bh_RDR) gtInt_RDR tag_lit) (ltResult op) $ -- Definitely LT nlHsCase (nlHsVar b_RDR) [ mkInnerEqAlt op data_con , mkHsCaseAlt nlWildPat (gtResult op) ] where tag = get_tag data_con tag_lit = noLoc (HsLit (HsIntPrim "" (toInteger tag))) mkInnerEqAlt :: OrdOp -> DataCon -> LMatch RdrName (LHsExpr RdrName) -- First argument 'a' known to be built with K -- Returns a case alternative Ki b1 b2 ... bv -> compare (a1,a2,...) with (b1,b2,...) mkInnerEqAlt op data_con = mkHsCaseAlt (nlConVarPat data_con_RDR bs_needed) $ mkCompareFields tycon op (dataConOrigArgTys data_con) where data_con_RDR = getRdrName data_con bs_needed = take (dataConSourceArity data_con) bs_RDRs mkTagCmp :: OrdOp -> LHsExpr RdrName -- Both constructors known to be nullary -- genreates (case data2Tag a of a# -> case data2Tag b of b# -> a# `op` b# mkTagCmp op = untag_Expr tycon [(a_RDR, ah_RDR),(b_RDR, bh_RDR)] $ unliftedOrdOp tycon intPrimTy op ah_RDR bh_RDR mkCompareFields :: TyCon -> OrdOp -> [Type] -> LHsExpr RdrName -- Generates nested comparisons for (a1,a2...) against (b1,b2,...) -- where the ai,bi have the given types mkCompareFields tycon op tys = go tys as_RDRs bs_RDRs where go [] _ _ = eqResult op go [ty] (a:_) (b:_) | isUnliftedType ty = unliftedOrdOp tycon ty op a b | otherwise = genOpApp (nlHsVar a) (ordMethRdr op) (nlHsVar b) go (ty:tys) (a:as) (b:bs) = mk_compare ty a b (ltResult op) (go tys as bs) (gtResult op) go _ _ _ = panic "mkCompareFields" -- (mk_compare ty a b) generates -- (case (compare a b) of { LT -> <lt>; EQ -> <eq>; GT -> <bt> }) -- but with suitable special cases for mk_compare ty a b lt eq gt | isUnliftedType ty = unliftedCompare lt_op eq_op a_expr b_expr lt eq gt | otherwise = nlHsCase (nlHsPar (nlHsApp (nlHsApp (nlHsVar compare_RDR) a_expr) b_expr)) [mkHsCaseAlt (nlNullaryConPat ltTag_RDR) lt, mkHsCaseAlt (nlNullaryConPat eqTag_RDR) eq, mkHsCaseAlt (nlNullaryConPat gtTag_RDR) gt] where a_expr = nlHsVar a b_expr = nlHsVar b (lt_op, _, eq_op, _, _) = primOrdOps "Ord" tycon ty unliftedOrdOp :: TyCon -> Type -> OrdOp -> RdrName -> RdrName -> LHsExpr RdrName unliftedOrdOp tycon ty op a b = case op of OrdCompare -> unliftedCompare lt_op eq_op a_expr b_expr ltTag_Expr eqTag_Expr gtTag_Expr OrdLT -> wrap lt_op OrdLE -> wrap le_op OrdGE -> wrap ge_op OrdGT -> wrap gt_op where (lt_op, le_op, eq_op, ge_op, gt_op) = primOrdOps "Ord" tycon ty wrap prim_op = genPrimOpApp a_expr prim_op b_expr a_expr = nlHsVar a b_expr = nlHsVar b unliftedCompare :: RdrName -> RdrName -> LHsExpr RdrName -> LHsExpr RdrName -- What to cmpare -> LHsExpr RdrName -> LHsExpr RdrName -> LHsExpr RdrName -- Three results -> LHsExpr RdrName -- Return (if a < b then lt else if a == b then eq else gt) unliftedCompare lt_op eq_op a_expr b_expr lt eq gt = nlHsIf (ascribeBool $ genPrimOpApp a_expr lt_op b_expr) lt $ -- Test (<) first, not (==), because the latter -- is true less often, so putting it first would -- mean more tests (dynamically) nlHsIf (ascribeBool $ genPrimOpApp a_expr eq_op b_expr) eq gt where ascribeBool e = nlExprWithTySig e (toLHsSigWcType boolTy) nlConWildPat :: DataCon -> LPat RdrName -- The pattern (K {}) nlConWildPat con = noLoc (ConPatIn (noLoc (getRdrName con)) (RecCon (HsRecFields { rec_flds = [] , rec_dotdot = Nothing }))) {- ************************************************************************ * * Enum instances * * ************************************************************************ @Enum@ can only be derived for enumeration types. For a type \begin{verbatim} data Foo ... = N1 | N2 | ... | Nn \end{verbatim} we use both @con2tag_Foo@ and @tag2con_Foo@ functions, as well as a @maxtag_Foo@ variable (all generated by @gen_tag_n_con_binds@). \begin{verbatim} instance ... Enum (Foo ...) where succ x = toEnum (1 + fromEnum x) pred x = toEnum (fromEnum x - 1) toEnum i = tag2con_Foo i enumFrom a = map tag2con_Foo [con2tag_Foo a .. maxtag_Foo] -- or, really... enumFrom a = case con2tag_Foo a of a# -> map tag2con_Foo (enumFromTo (I# a#) maxtag_Foo) enumFromThen a b = map tag2con_Foo [con2tag_Foo a, con2tag_Foo b .. maxtag_Foo] -- or, really... enumFromThen a b = case con2tag_Foo a of { a# -> case con2tag_Foo b of { b# -> map tag2con_Foo (enumFromThenTo (I# a#) (I# b#) maxtag_Foo) }} \end{verbatim} For @enumFromTo@ and @enumFromThenTo@, we use the default methods. -} gen_Enum_binds :: SrcSpan -> TyCon -> (LHsBinds RdrName, BagDerivStuff) gen_Enum_binds loc tycon = (method_binds, aux_binds) where method_binds = listToBag [ succ_enum, pred_enum, to_enum, enum_from, enum_from_then, from_enum ] aux_binds = listToBag $ map DerivAuxBind [DerivCon2Tag tycon, DerivTag2Con tycon, DerivMaxTag tycon] occ_nm = getOccString tycon succ_enum = mk_easy_FunBind loc succ_RDR [a_Pat] $ untag_Expr tycon [(a_RDR, ah_RDR)] $ nlHsIf (nlHsApps eq_RDR [nlHsVar (maxtag_RDR tycon), nlHsVarApps intDataCon_RDR [ah_RDR]]) (illegal_Expr "succ" occ_nm "tried to take `succ' of last tag in enumeration") (nlHsApp (nlHsVar (tag2con_RDR tycon)) (nlHsApps plus_RDR [nlHsVarApps intDataCon_RDR [ah_RDR], nlHsIntLit 1])) pred_enum = mk_easy_FunBind loc pred_RDR [a_Pat] $ untag_Expr tycon [(a_RDR, ah_RDR)] $ nlHsIf (nlHsApps eq_RDR [nlHsIntLit 0, nlHsVarApps intDataCon_RDR [ah_RDR]]) (illegal_Expr "pred" occ_nm "tried to take `pred' of first tag in enumeration") (nlHsApp (nlHsVar (tag2con_RDR tycon)) (nlHsApps plus_RDR [nlHsVarApps intDataCon_RDR [ah_RDR], nlHsLit (HsInt "-1" (-1))])) to_enum = mk_easy_FunBind loc toEnum_RDR [a_Pat] $ nlHsIf (nlHsApps and_RDR [nlHsApps ge_RDR [nlHsVar a_RDR, nlHsIntLit 0], nlHsApps le_RDR [nlHsVar a_RDR, nlHsVar (maxtag_RDR tycon)]]) (nlHsVarApps (tag2con_RDR tycon) [a_RDR]) (illegal_toEnum_tag occ_nm (maxtag_RDR tycon)) enum_from = mk_easy_FunBind loc enumFrom_RDR [a_Pat] $ untag_Expr tycon [(a_RDR, ah_RDR)] $ nlHsApps map_RDR [nlHsVar (tag2con_RDR tycon), nlHsPar (enum_from_to_Expr (nlHsVarApps intDataCon_RDR [ah_RDR]) (nlHsVar (maxtag_RDR tycon)))] enum_from_then = mk_easy_FunBind loc enumFromThen_RDR [a_Pat, b_Pat] $ untag_Expr tycon [(a_RDR, ah_RDR), (b_RDR, bh_RDR)] $ nlHsApp (nlHsVarApps map_RDR [tag2con_RDR tycon]) $ nlHsPar (enum_from_then_to_Expr (nlHsVarApps intDataCon_RDR [ah_RDR]) (nlHsVarApps intDataCon_RDR [bh_RDR]) (nlHsIf (nlHsApps gt_RDR [nlHsVarApps intDataCon_RDR [ah_RDR], nlHsVarApps intDataCon_RDR [bh_RDR]]) (nlHsIntLit 0) (nlHsVar (maxtag_RDR tycon)) )) from_enum = mk_easy_FunBind loc fromEnum_RDR [a_Pat] $ untag_Expr tycon [(a_RDR, ah_RDR)] $ (nlHsVarApps intDataCon_RDR [ah_RDR]) {- ************************************************************************ * * Bounded instances * * ************************************************************************ -} gen_Bounded_binds :: SrcSpan -> TyCon -> (LHsBinds RdrName, BagDerivStuff) gen_Bounded_binds loc tycon | isEnumerationTyCon tycon = (listToBag [ min_bound_enum, max_bound_enum ], emptyBag) | otherwise = ASSERT(isSingleton data_cons) (listToBag [ min_bound_1con, max_bound_1con ], emptyBag) where data_cons = tyConDataCons tycon ----- enum-flavored: --------------------------- min_bound_enum = mkHsVarBind loc minBound_RDR (nlHsVar data_con_1_RDR) max_bound_enum = mkHsVarBind loc maxBound_RDR (nlHsVar data_con_N_RDR) data_con_1 = head data_cons data_con_N = last data_cons data_con_1_RDR = getRdrName data_con_1 data_con_N_RDR = getRdrName data_con_N ----- single-constructor-flavored: ------------- arity = dataConSourceArity data_con_1 min_bound_1con = mkHsVarBind loc minBound_RDR $ nlHsVarApps data_con_1_RDR (nOfThem arity minBound_RDR) max_bound_1con = mkHsVarBind loc maxBound_RDR $ nlHsVarApps data_con_1_RDR (nOfThem arity maxBound_RDR) {- ************************************************************************ * * Ix instances * * ************************************************************************ Deriving @Ix@ is only possible for enumeration types and single-constructor types. We deal with them in turn. For an enumeration type, e.g., \begin{verbatim} data Foo ... = N1 | N2 | ... | Nn \end{verbatim} things go not too differently from @Enum@: \begin{verbatim} instance ... Ix (Foo ...) where range (a, b) = map tag2con_Foo [con2tag_Foo a .. con2tag_Foo b] -- or, really... range (a, b) = case (con2tag_Foo a) of { a# -> case (con2tag_Foo b) of { b# -> map tag2con_Foo (enumFromTo (I# a#) (I# b#)) }} -- Generate code for unsafeIndex, because using index leads -- to lots of redundant range tests unsafeIndex c@(a, b) d = case (con2tag_Foo d -# con2tag_Foo a) of r# -> I# r# inRange (a, b) c = let p_tag = con2tag_Foo c in p_tag >= con2tag_Foo a && p_tag <= con2tag_Foo b -- or, really... inRange (a, b) c = case (con2tag_Foo a) of { a_tag -> case (con2tag_Foo b) of { b_tag -> case (con2tag_Foo c) of { c_tag -> if (c_tag >=# a_tag) then c_tag <=# b_tag else False }}} \end{verbatim} (modulo suitable case-ification to handle the unlifted tags) For a single-constructor type (NB: this includes all tuples), e.g., \begin{verbatim} data Foo ... = MkFoo a b Int Double c c \end{verbatim} we follow the scheme given in Figure~19 of the Haskell~1.2 report (p.~147). -} gen_Ix_binds :: SrcSpan -> TyCon -> (LHsBinds RdrName, BagDerivStuff) gen_Ix_binds loc tycon | isEnumerationTyCon tycon = ( enum_ixes , listToBag $ map DerivAuxBind [DerivCon2Tag tycon, DerivTag2Con tycon, DerivMaxTag tycon]) | otherwise = (single_con_ixes, unitBag (DerivAuxBind (DerivCon2Tag tycon))) where -------------------------------------------------------------- enum_ixes = listToBag [ enum_range, enum_index, enum_inRange ] enum_range = mk_easy_FunBind loc range_RDR [nlTuplePat [a_Pat, b_Pat] Boxed] $ untag_Expr tycon [(a_RDR, ah_RDR)] $ untag_Expr tycon [(b_RDR, bh_RDR)] $ nlHsApp (nlHsVarApps map_RDR [tag2con_RDR tycon]) $ nlHsPar (enum_from_to_Expr (nlHsVarApps intDataCon_RDR [ah_RDR]) (nlHsVarApps intDataCon_RDR [bh_RDR])) enum_index = mk_easy_FunBind loc unsafeIndex_RDR [noLoc (AsPat (noLoc c_RDR) (nlTuplePat [a_Pat, nlWildPat] Boxed)), d_Pat] ( untag_Expr tycon [(a_RDR, ah_RDR)] ( untag_Expr tycon [(d_RDR, dh_RDR)] ( let rhs = nlHsVarApps intDataCon_RDR [c_RDR] in nlHsCase (genOpApp (nlHsVar dh_RDR) minusInt_RDR (nlHsVar ah_RDR)) [mkHsCaseAlt (nlVarPat c_RDR) rhs] )) ) -- This produces something like `(ch >= ah) && (ch <= bh)` enum_inRange = mk_easy_FunBind loc inRange_RDR [nlTuplePat [a_Pat, b_Pat] Boxed, c_Pat] $ untag_Expr tycon [(a_RDR, ah_RDR)] ( untag_Expr tycon [(b_RDR, bh_RDR)] ( untag_Expr tycon [(c_RDR, ch_RDR)] ( -- This used to use `if`, which interacts badly with RebindableSyntax. -- See #11396. nlHsApps and_RDR [ genPrimOpApp (nlHsVar ch_RDR) geInt_RDR (nlHsVar ah_RDR) , genPrimOpApp (nlHsVar ch_RDR) leInt_RDR (nlHsVar bh_RDR) ] ))) -------------------------------------------------------------- single_con_ixes = listToBag [single_con_range, single_con_index, single_con_inRange] data_con = case tyConSingleDataCon_maybe tycon of -- just checking... Nothing -> panic "get_Ix_binds" Just dc -> dc con_arity = dataConSourceArity data_con data_con_RDR = getRdrName data_con as_needed = take con_arity as_RDRs bs_needed = take con_arity bs_RDRs cs_needed = take con_arity cs_RDRs con_pat xs = nlConVarPat data_con_RDR xs con_expr = nlHsVarApps data_con_RDR cs_needed -------------------------------------------------------------- single_con_range = mk_easy_FunBind loc range_RDR [nlTuplePat [con_pat as_needed, con_pat bs_needed] Boxed] $ noLoc (mkHsComp ListComp stmts con_expr) where stmts = zipWith3Equal "single_con_range" mk_qual as_needed bs_needed cs_needed mk_qual a b c = noLoc $ mkBindStmt (nlVarPat c) (nlHsApp (nlHsVar range_RDR) (mkLHsVarTuple [a,b])) ---------------- single_con_index = mk_easy_FunBind loc unsafeIndex_RDR [nlTuplePat [con_pat as_needed, con_pat bs_needed] Boxed, con_pat cs_needed] -- We need to reverse the order we consider the components in -- so that -- range (l,u) !! index (l,u) i == i -- when i is in range -- (from http://haskell.org/onlinereport/ix.html) holds. (mk_index (reverse $ zip3 as_needed bs_needed cs_needed)) where -- index (l1,u1) i1 + rangeSize (l1,u1) * (index (l2,u2) i2 + ...) mk_index [] = nlHsIntLit 0 mk_index [(l,u,i)] = mk_one l u i mk_index ((l,u,i) : rest) = genOpApp ( mk_one l u i ) plus_RDR ( genOpApp ( (nlHsApp (nlHsVar unsafeRangeSize_RDR) (mkLHsVarTuple [l,u])) ) times_RDR (mk_index rest) ) mk_one l u i = nlHsApps unsafeIndex_RDR [mkLHsVarTuple [l,u], nlHsVar i] ------------------ single_con_inRange = mk_easy_FunBind loc inRange_RDR [nlTuplePat [con_pat as_needed, con_pat bs_needed] Boxed, con_pat cs_needed] $ foldl1 and_Expr (zipWith3Equal "single_con_inRange" in_range as_needed bs_needed cs_needed) where in_range a b c = nlHsApps inRange_RDR [mkLHsVarTuple [a,b], nlHsVar c] {- ************************************************************************ * * Read instances * * ************************************************************************ Example infix 4 %% data T = Int %% Int | T1 { f1 :: Int } | T2 T instance Read T where readPrec = parens ( prec 4 ( do x <- ReadP.step Read.readPrec expectP (Symbol "%%") y <- ReadP.step Read.readPrec return (x %% y)) +++ prec (appPrec+1) ( -- Note the "+1" part; "T2 T1 {f1=3}" should parse ok -- Record construction binds even more tightly than application do expectP (Ident "T1") expectP (Punc '{') expectP (Ident "f1") expectP (Punc '=') x <- ReadP.reset Read.readPrec expectP (Punc '}') return (T1 { f1 = x })) +++ prec appPrec ( do expectP (Ident "T2") x <- ReadP.step Read.readPrec return (T2 x)) ) readListPrec = readListPrecDefault readList = readListDefault Note [Use expectP] ~~~~~~~~~~~~~~~~~~ Note that we use expectP (Ident "T1") rather than Ident "T1" <- lexP The latter desugares to inline code for matching the Ident and the string, and this can be very voluminous. The former is much more compact. Cf Trac #7258, although that also concerned non-linearity in the occurrence analyser, a separate issue. Note [Read for empty data types] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ What should we get for this? (Trac #7931) data Emp deriving( Read ) -- No data constructors Here we want read "[]" :: [Emp] to succeed, returning [] So we do NOT want instance Read Emp where readPrec = error "urk" Rather we want instance Read Emp where readPred = pfail -- Same as choose [] Because 'pfail' allows the parser to backtrack, but 'error' doesn't. These instances are also useful for Read (Either Int Emp), where we want to be able to parse (Left 3) just fine. -} gen_Read_binds :: (Name -> Fixity) -> SrcSpan -> TyCon -> (LHsBinds RdrName, BagDerivStuff) gen_Read_binds get_fixity loc tycon = (listToBag [read_prec, default_readlist, default_readlistprec], emptyBag) where ----------------------------------------------------------------------- default_readlist = mkHsVarBind loc readList_RDR (nlHsVar readListDefault_RDR) default_readlistprec = mkHsVarBind loc readListPrec_RDR (nlHsVar readListPrecDefault_RDR) ----------------------------------------------------------------------- data_cons = tyConDataCons tycon (nullary_cons, non_nullary_cons) = partition isNullarySrcDataCon data_cons read_prec = mkHsVarBind loc readPrec_RDR (nlHsApp (nlHsVar parens_RDR) read_cons) read_cons | null data_cons = nlHsVar pfail_RDR -- See Note [Read for empty data types] | otherwise = foldr1 mk_alt (read_nullary_cons ++ read_non_nullary_cons) read_non_nullary_cons = map read_non_nullary_con non_nullary_cons read_nullary_cons = case nullary_cons of [] -> [] [con] -> [nlHsDo DoExpr (match_con con ++ [noLoc $ mkLastStmt (result_expr con [])])] _ -> [nlHsApp (nlHsVar choose_RDR) (nlList (map mk_pair nullary_cons))] -- NB For operators the parens around (:=:) are matched by the -- enclosing "parens" call, so here we must match the naked -- data_con_str con match_con con | isSym con_str = [symbol_pat con_str] | otherwise = ident_h_pat con_str where con_str = data_con_str con -- For nullary constructors we must match Ident s for normal constrs -- and Symbol s for operators mk_pair con = mkLHsTupleExpr [nlHsLit (mkHsString (data_con_str con)), result_expr con []] read_non_nullary_con data_con | is_infix = mk_parser infix_prec infix_stmts body | is_record = mk_parser record_prec record_stmts body -- Using these two lines instead allows the derived -- read for infix and record bindings to read the prefix form -- | is_infix = mk_alt prefix_parser (mk_parser infix_prec infix_stmts body) -- | is_record = mk_alt prefix_parser (mk_parser record_prec record_stmts body) | otherwise = prefix_parser where body = result_expr data_con as_needed con_str = data_con_str data_con prefix_parser = mk_parser prefix_prec prefix_stmts body read_prefix_con | isSym con_str = [read_punc "(", symbol_pat con_str, read_punc ")"] | otherwise = ident_h_pat con_str read_infix_con | isSym con_str = [symbol_pat con_str] | otherwise = [read_punc "`"] ++ ident_h_pat con_str ++ [read_punc "`"] prefix_stmts -- T a b c = read_prefix_con ++ read_args infix_stmts -- a %% b, or a `T` b = [read_a1] ++ read_infix_con ++ [read_a2] record_stmts -- T { f1 = a, f2 = b } = read_prefix_con ++ [read_punc "{"] ++ concat (intersperse [read_punc ","] field_stmts) ++ [read_punc "}"] field_stmts = zipWithEqual "lbl_stmts" read_field labels as_needed con_arity = dataConSourceArity data_con labels = map flLabel $ dataConFieldLabels data_con dc_nm = getName data_con is_infix = dataConIsInfix data_con is_record = length labels > 0 as_needed = take con_arity as_RDRs read_args = zipWithEqual "gen_Read_binds" read_arg as_needed (dataConOrigArgTys data_con) (read_a1:read_a2:_) = read_args prefix_prec = appPrecedence infix_prec = getPrecedence get_fixity dc_nm record_prec = appPrecedence + 1 -- Record construction binds even more tightly -- than application; e.g. T2 T1 {x=2} means T2 (T1 {x=2}) ------------------------------------------------------------------------ -- Helpers ------------------------------------------------------------------------ mk_alt e1 e2 = genOpApp e1 alt_RDR e2 -- e1 +++ e2 mk_parser p ss b = nlHsApps prec_RDR [nlHsIntLit p -- prec p (do { ss ; b }) , nlHsDo DoExpr (ss ++ [noLoc $ mkLastStmt b])] con_app con as = nlHsVarApps (getRdrName con) as -- con as result_expr con as = nlHsApp (nlHsVar returnM_RDR) (con_app con as) -- return (con as) -- For constructors and field labels ending in '#', we hackily -- let the lexer generate two tokens, and look for both in sequence -- Thus [Ident "I"; Symbol "#"]. See Trac #5041 ident_h_pat s | Just (ss, '#') <- snocView s = [ ident_pat ss, symbol_pat "#" ] | otherwise = [ ident_pat s ] bindLex pat = noLoc (mkBodyStmt (nlHsApp (nlHsVar expectP_RDR) pat)) -- expectP p -- See Note [Use expectP] ident_pat s = bindLex $ nlHsApps ident_RDR [nlHsLit (mkHsString s)] -- expectP (Ident "foo") symbol_pat s = bindLex $ nlHsApps symbol_RDR [nlHsLit (mkHsString s)] -- expectP (Symbol ">>") read_punc c = bindLex $ nlHsApps punc_RDR [nlHsLit (mkHsString c)] -- expectP (Punc "<") data_con_str con = occNameString (getOccName con) read_arg a ty = ASSERT( not (isUnliftedType ty) ) noLoc (mkBindStmt (nlVarPat a) (nlHsVarApps step_RDR [readPrec_RDR])) read_field lbl a = read_lbl lbl ++ [read_punc "=", noLoc (mkBindStmt (nlVarPat a) (nlHsVarApps reset_RDR [readPrec_RDR]))] -- When reading field labels we might encounter -- a = 3 -- _a = 3 -- or (#) = 4 -- Note the parens! read_lbl lbl | isSym lbl_str = [read_punc "(", symbol_pat lbl_str, read_punc ")"] | otherwise = ident_h_pat lbl_str where lbl_str = unpackFS lbl {- ************************************************************************ * * Show instances * * ************************************************************************ Example infixr 5 :^: data Tree a = Leaf a | Tree a :^: Tree a instance (Show a) => Show (Tree a) where showsPrec d (Leaf m) = showParen (d > app_prec) showStr where showStr = showString "Leaf " . showsPrec (app_prec+1) m showsPrec d (u :^: v) = showParen (d > up_prec) showStr where showStr = showsPrec (up_prec+1) u . showString " :^: " . showsPrec (up_prec+1) v -- Note: right-associativity of :^: ignored up_prec = 5 -- Precedence of :^: app_prec = 10 -- Application has precedence one more than -- the most tightly-binding operator -} gen_Show_binds :: (Name -> Fixity) -> SrcSpan -> TyCon -> (LHsBinds RdrName, BagDerivStuff) gen_Show_binds get_fixity loc tycon = (listToBag [shows_prec, show_list], emptyBag) where ----------------------------------------------------------------------- show_list = mkHsVarBind loc showList_RDR (nlHsApp (nlHsVar showList___RDR) (nlHsPar (nlHsApp (nlHsVar showsPrec_RDR) (nlHsIntLit 0)))) ----------------------------------------------------------------------- data_cons = tyConDataCons tycon shows_prec = mk_FunBind loc showsPrec_RDR (map pats_etc data_cons) pats_etc data_con | nullary_con = -- skip the showParen junk... ASSERT(null bs_needed) ([nlWildPat, con_pat], mk_showString_app op_con_str) | otherwise = ([a_Pat, con_pat], showParen_Expr (genOpApp a_Expr ge_RDR (nlHsLit (HsInt "" con_prec_plus_one))) (nlHsPar (nested_compose_Expr show_thingies))) where data_con_RDR = getRdrName data_con con_arity = dataConSourceArity data_con bs_needed = take con_arity bs_RDRs arg_tys = dataConOrigArgTys data_con -- Correspond 1-1 with bs_needed con_pat = nlConVarPat data_con_RDR bs_needed nullary_con = con_arity == 0 labels = map flLabel $ dataConFieldLabels data_con lab_fields = length labels record_syntax = lab_fields > 0 dc_nm = getName data_con dc_occ_nm = getOccName data_con con_str = occNameString dc_occ_nm op_con_str = wrapOpParens con_str backquote_str = wrapOpBackquotes con_str show_thingies | is_infix = [show_arg1, mk_showString_app (" " ++ backquote_str ++ " "), show_arg2] | record_syntax = mk_showString_app (op_con_str ++ " {") : show_record_args ++ [mk_showString_app "}"] | otherwise = mk_showString_app (op_con_str ++ " ") : show_prefix_args show_label l = mk_showString_app (nm ++ " = ") -- Note the spaces around the "=" sign. If we -- don't have them then we get Foo { x=-1 } and -- the "=-" parses as a single lexeme. Only the -- space after the '=' is necessary, but it -- seems tidier to have them both sides. where nm = wrapOpParens (unpackFS l) show_args = zipWith show_arg bs_needed arg_tys (show_arg1:show_arg2:_) = show_args show_prefix_args = intersperse (nlHsVar showSpace_RDR) show_args -- Assumption for record syntax: no of fields == no of -- labelled fields (and in same order) show_record_args = concat $ intersperse [mk_showString_app ", "] $ [ [show_label lbl, arg] | (lbl,arg) <- zipEqual "gen_Show_binds" labels show_args ] show_arg :: RdrName -> Type -> LHsExpr RdrName show_arg b arg_ty | isUnliftedType arg_ty -- See Note [Deriving and unboxed types] in TcDeriv = nlHsApps compose_RDR [mk_shows_app boxed_arg, mk_showString_app postfixMod] | otherwise = mk_showsPrec_app arg_prec arg where arg = nlHsVar b boxed_arg = box "Show" tycon arg arg_ty postfixMod = assoc_ty_id "Show" tycon postfixModTbl arg_ty -- Fixity stuff is_infix = dataConIsInfix data_con con_prec_plus_one = 1 + getPrec is_infix get_fixity dc_nm arg_prec | record_syntax = 0 -- Record fields don't need parens | otherwise = con_prec_plus_one wrapOpParens :: String -> String wrapOpParens s | isSym s = '(' : s ++ ")" | otherwise = s wrapOpBackquotes :: String -> String wrapOpBackquotes s | isSym s = s | otherwise = '`' : s ++ "`" isSym :: String -> Bool isSym "" = False isSym (c : _) = startsVarSym c || startsConSym c -- | showString :: String -> ShowS mk_showString_app :: String -> LHsExpr RdrName mk_showString_app str = nlHsApp (nlHsVar showString_RDR) (nlHsLit (mkHsString str)) -- | showsPrec :: Show a => Int -> a -> ShowS mk_showsPrec_app :: Integer -> LHsExpr RdrName -> LHsExpr RdrName mk_showsPrec_app p x = nlHsApps showsPrec_RDR [nlHsLit (HsInt "" p), x] -- | shows :: Show a => a -> ShowS mk_shows_app :: LHsExpr RdrName -> LHsExpr RdrName mk_shows_app x = nlHsApp (nlHsVar shows_RDR) x getPrec :: Bool -> (Name -> Fixity) -> Name -> Integer getPrec is_infix get_fixity nm | not is_infix = appPrecedence | otherwise = getPrecedence get_fixity nm appPrecedence :: Integer appPrecedence = fromIntegral maxPrecedence + 1 -- One more than the precedence of the most -- tightly-binding operator getPrecedence :: (Name -> Fixity) -> Name -> Integer getPrecedence get_fixity nm = case get_fixity nm of Fixity _ x _assoc -> fromIntegral x -- NB: the Report says that associativity is not taken -- into account for either Read or Show; hence we -- ignore associativity here {- ************************************************************************ * * Data instances * * ************************************************************************ From the data type data T a b = T1 a b | T2 we generate $cT1 = mkDataCon $dT "T1" Prefix $cT2 = mkDataCon $dT "T2" Prefix $dT = mkDataType "Module.T" [] [$con_T1, $con_T2] -- the [] is for field labels. instance (Data a, Data b) => Data (T a b) where gfoldl k z (T1 a b) = z T `k` a `k` b gfoldl k z T2 = z T2 -- ToDo: add gmapT,Q,M, gfoldr gunfold k z c = case conIndex c of I# 1# -> k (k (z T1)) I# 2# -> z T2 toConstr (T1 _ _) = $cT1 toConstr T2 = $cT2 dataTypeOf _ = $dT dataCast1 = gcast1 -- If T :: * -> * dataCast2 = gcast2 -- if T :: * -> * -> * -} gen_Data_binds :: SrcSpan -> TyCon -- For data families, this is the -- *representation* TyCon -> TcM (LHsBinds RdrName, -- The method bindings BagDerivStuff) -- Auxiliary bindings gen_Data_binds loc rep_tc = do { dflags <- getDynFlags -- Make unique names for the data type and constructor -- auxiliary bindings. Start with the name of the TyCon/DataCon -- but that might not be unique: see Trac #12245. ; dt_occ <- chooseUniqueOccTc (mkDataTOcc (getOccName rep_tc)) ; dc_occs <- mapM (chooseUniqueOccTc . mkDataCOcc . getOccName) (tyConDataCons rep_tc) ; let dt_rdr = mkRdrUnqual dt_occ dc_rdrs = map mkRdrUnqual dc_occs -- OK, now do the work ; return (gen_data dflags dt_rdr dc_rdrs loc rep_tc) } gen_data :: DynFlags -> RdrName -> [RdrName] -> SrcSpan -> TyCon -> (LHsBinds RdrName, -- The method bindings BagDerivStuff) -- Auxiliary bindings gen_data dflags data_type_name constr_names loc rep_tc = (listToBag [gfoldl_bind, gunfold_bind, toCon_bind, dataTypeOf_bind] `unionBags` gcast_binds, -- Auxiliary definitions: the data type and constructors listToBag ( genDataTyCon : zipWith genDataDataCon data_cons constr_names ) ) where data_cons = tyConDataCons rep_tc n_cons = length data_cons one_constr = n_cons == 1 genDataTyCon :: DerivStuff genDataTyCon -- $dT = DerivHsBind (mkHsVarBind loc data_type_name rhs, L loc (TypeSig [L loc data_type_name] sig_ty)) sig_ty = mkLHsSigWcType (nlHsTyVar dataType_RDR) rhs = nlHsVar mkDataType_RDR `nlHsApp` nlHsLit (mkHsString (showSDocOneLine dflags (ppr rep_tc))) `nlHsApp` nlList (map nlHsVar constr_names) genDataDataCon :: DataCon -> RdrName -> DerivStuff genDataDataCon dc constr_name -- $cT1 etc = DerivHsBind (mkHsVarBind loc constr_name rhs, L loc (TypeSig [L loc constr_name] sig_ty)) where sig_ty = mkLHsSigWcType (nlHsTyVar constr_RDR) rhs = nlHsApps mkConstr_RDR constr_args constr_args = [ -- nlHsIntLit (toInteger (dataConTag dc)), -- Tag nlHsVar (data_type_name) -- DataType , nlHsLit (mkHsString (occNameString dc_occ)) -- String name , nlList labels -- Field labels , nlHsVar fixity ] -- Fixity labels = map (nlHsLit . mkHsString . unpackFS . flLabel) (dataConFieldLabels dc) dc_occ = getOccName dc is_infix = isDataSymOcc dc_occ fixity | is_infix = infix_RDR | otherwise = prefix_RDR ------------ gfoldl gfoldl_bind = mk_HRFunBind 2 loc gfoldl_RDR (map gfoldl_eqn data_cons) gfoldl_eqn con = ([nlVarPat k_RDR, nlVarPat z_RDR, nlConVarPat con_name as_needed], foldl mk_k_app (nlHsVar z_RDR `nlHsApp` nlHsVar con_name) as_needed) where con_name :: RdrName con_name = getRdrName con as_needed = take (dataConSourceArity con) as_RDRs mk_k_app e v = nlHsPar (nlHsOpApp e k_RDR (nlHsVar v)) ------------ gunfold gunfold_bind = mk_HRFunBind 2 loc gunfold_RDR [([k_Pat, z_Pat, if one_constr then nlWildPat else c_Pat], gunfold_rhs)] gunfold_rhs | one_constr = mk_unfold_rhs (head data_cons) -- No need for case | otherwise = nlHsCase (nlHsVar conIndex_RDR `nlHsApp` c_Expr) (map gunfold_alt data_cons) gunfold_alt dc = mkHsCaseAlt (mk_unfold_pat dc) (mk_unfold_rhs dc) mk_unfold_rhs dc = foldr nlHsApp (nlHsVar z_RDR `nlHsApp` nlHsVar (getRdrName dc)) (replicate (dataConSourceArity dc) (nlHsVar k_RDR)) mk_unfold_pat dc -- Last one is a wild-pat, to avoid -- redundant test, and annoying warning | tag-fIRST_TAG == n_cons-1 = nlWildPat -- Last constructor | otherwise = nlConPat intDataCon_RDR [nlLitPat (HsIntPrim "" (toInteger tag))] where tag = dataConTag dc ------------ toConstr toCon_bind = mk_FunBind loc toConstr_RDR (zipWith to_con_eqn data_cons constr_names) to_con_eqn dc con_name = ([nlWildConPat dc], nlHsVar con_name) ------------ dataTypeOf dataTypeOf_bind = mk_easy_FunBind loc dataTypeOf_RDR [nlWildPat] (nlHsVar data_type_name) ------------ gcast1/2 -- Make the binding dataCast1 x = gcast1 x -- if T :: * -> * -- or dataCast2 x = gcast2 s -- if T :: * -> * -> * -- (or nothing if T has neither of these two types) -- But care is needed for data families: -- If we have data family D a -- data instance D (a,b,c) = A | B deriving( Data ) -- and we want instance ... => Data (D [(a,b,c)]) where ... -- then we need dataCast1 x = gcast1 x -- because D :: * -> * -- even though rep_tc has kind * -> * -> * -> * -- Hence looking for the kind of fam_tc not rep_tc -- See Trac #4896 tycon_kind = case tyConFamInst_maybe rep_tc of Just (fam_tc, _) -> tyConKind fam_tc Nothing -> tyConKind rep_tc gcast_binds | tycon_kind `tcEqKind` kind1 = mk_gcast dataCast1_RDR gcast1_RDR | tycon_kind `tcEqKind` kind2 = mk_gcast dataCast2_RDR gcast2_RDR | otherwise = emptyBag mk_gcast dataCast_RDR gcast_RDR = unitBag (mk_easy_FunBind loc dataCast_RDR [nlVarPat f_RDR] (nlHsVar gcast_RDR `nlHsApp` nlHsVar f_RDR)) kind1, kind2 :: Kind kind1 = liftedTypeKind `mkFunTy` liftedTypeKind kind2 = liftedTypeKind `mkFunTy` kind1 gfoldl_RDR, gunfold_RDR, toConstr_RDR, dataTypeOf_RDR, mkConstr_RDR, mkDataType_RDR, conIndex_RDR, prefix_RDR, infix_RDR, dataCast1_RDR, dataCast2_RDR, gcast1_RDR, gcast2_RDR, constr_RDR, dataType_RDR, eqChar_RDR , ltChar_RDR , geChar_RDR , gtChar_RDR , leChar_RDR , eqInt_RDR , ltInt_RDR , geInt_RDR , gtInt_RDR , leInt_RDR , eqWord_RDR , ltWord_RDR , geWord_RDR , gtWord_RDR , leWord_RDR , eqAddr_RDR , ltAddr_RDR , geAddr_RDR , gtAddr_RDR , leAddr_RDR , eqFloat_RDR , ltFloat_RDR , geFloat_RDR , gtFloat_RDR , leFloat_RDR , eqDouble_RDR, ltDouble_RDR, geDouble_RDR, gtDouble_RDR, leDouble_RDR :: RdrName gfoldl_RDR = varQual_RDR gENERICS (fsLit "gfoldl") gunfold_RDR = varQual_RDR gENERICS (fsLit "gunfold") toConstr_RDR = varQual_RDR gENERICS (fsLit "toConstr") dataTypeOf_RDR = varQual_RDR gENERICS (fsLit "dataTypeOf") dataCast1_RDR = varQual_RDR gENERICS (fsLit "dataCast1") dataCast2_RDR = varQual_RDR gENERICS (fsLit "dataCast2") gcast1_RDR = varQual_RDR tYPEABLE (fsLit "gcast1") gcast2_RDR = varQual_RDR tYPEABLE (fsLit "gcast2") mkConstr_RDR = varQual_RDR gENERICS (fsLit "mkConstr") constr_RDR = tcQual_RDR gENERICS (fsLit "Constr") mkDataType_RDR = varQual_RDR gENERICS (fsLit "mkDataType") dataType_RDR = tcQual_RDR gENERICS (fsLit "DataType") conIndex_RDR = varQual_RDR gENERICS (fsLit "constrIndex") prefix_RDR = dataQual_RDR gENERICS (fsLit "Prefix") infix_RDR = dataQual_RDR gENERICS (fsLit "Infix") eqChar_RDR = varQual_RDR gHC_PRIM (fsLit "eqChar#") ltChar_RDR = varQual_RDR gHC_PRIM (fsLit "ltChar#") leChar_RDR = varQual_RDR gHC_PRIM (fsLit "leChar#") gtChar_RDR = varQual_RDR gHC_PRIM (fsLit "gtChar#") geChar_RDR = varQual_RDR gHC_PRIM (fsLit "geChar#") eqInt_RDR = varQual_RDR gHC_PRIM (fsLit "==#") ltInt_RDR = varQual_RDR gHC_PRIM (fsLit "<#" ) leInt_RDR = varQual_RDR gHC_PRIM (fsLit "<=#") gtInt_RDR = varQual_RDR gHC_PRIM (fsLit ">#" ) geInt_RDR = varQual_RDR gHC_PRIM (fsLit ">=#") eqWord_RDR = varQual_RDR gHC_PRIM (fsLit "eqWord#") ltWord_RDR = varQual_RDR gHC_PRIM (fsLit "ltWord#") leWord_RDR = varQual_RDR gHC_PRIM (fsLit "leWord#") gtWord_RDR = varQual_RDR gHC_PRIM (fsLit "gtWord#") geWord_RDR = varQual_RDR gHC_PRIM (fsLit "geWord#") eqAddr_RDR = varQual_RDR gHC_PRIM (fsLit "eqAddr#") ltAddr_RDR = varQual_RDR gHC_PRIM (fsLit "ltAddr#") leAddr_RDR = varQual_RDR gHC_PRIM (fsLit "leAddr#") gtAddr_RDR = varQual_RDR gHC_PRIM (fsLit "gtAddr#") geAddr_RDR = varQual_RDR gHC_PRIM (fsLit "geAddr#") eqFloat_RDR = varQual_RDR gHC_PRIM (fsLit "eqFloat#") ltFloat_RDR = varQual_RDR gHC_PRIM (fsLit "ltFloat#") leFloat_RDR = varQual_RDR gHC_PRIM (fsLit "leFloat#") gtFloat_RDR = varQual_RDR gHC_PRIM (fsLit "gtFloat#") geFloat_RDR = varQual_RDR gHC_PRIM (fsLit "geFloat#") eqDouble_RDR = varQual_RDR gHC_PRIM (fsLit "==##") ltDouble_RDR = varQual_RDR gHC_PRIM (fsLit "<##" ) leDouble_RDR = varQual_RDR gHC_PRIM (fsLit "<=##") gtDouble_RDR = varQual_RDR gHC_PRIM (fsLit ">##" ) geDouble_RDR = varQual_RDR gHC_PRIM (fsLit ">=##") {- ************************************************************************ * * Functor instances see http://www.mail-archive.com/haskell-prime@haskell.org/msg02116.html * * ************************************************************************ For the data type: data T a = T1 Int a | T2 (T a) We generate the instance: instance Functor T where fmap f (T1 b1 a) = T1 b1 (f a) fmap f (T2 ta) = T2 (fmap f ta) Notice that we don't simply apply 'fmap' to the constructor arguments. Rather - Do nothing to an argument whose type doesn't mention 'a' - Apply 'f' to an argument of type 'a' - Apply 'fmap f' to other arguments That's why we have to recurse deeply into the constructor argument types, rather than just one level, as we typically do. What about types with more than one type parameter? In general, we only derive Functor for the last position: data S a b = S1 [b] | S2 (a, T a b) instance Functor (S a) where fmap f (S1 bs) = S1 (fmap f bs) fmap f (S2 (p,q)) = S2 (a, fmap f q) However, we have special cases for - tuples - functions More formally, we write the derivation of fmap code over type variable 'a for type 'b as ($fmap 'a 'b). In this general notation the derived instance for T is: instance Functor T where fmap f (T1 x1 x2) = T1 ($(fmap 'a 'b1) x1) ($(fmap 'a 'a) x2) fmap f (T2 x1) = T2 ($(fmap 'a '(T a)) x1) $(fmap 'a 'b) = \x -> x -- when b does not contain a $(fmap 'a 'a) = f $(fmap 'a '(b1,b2)) = \x -> case x of (x1,x2) -> ($(fmap 'a 'b1) x1, $(fmap 'a 'b2) x2) $(fmap 'a '(T b1 b2)) = fmap $(fmap 'a 'b2) -- when a only occurs in the last parameter, b2 $(fmap 'a '(b -> c)) = \x b -> $(fmap 'a' 'c) (x ($(cofmap 'a 'b) b)) For functions, the type parameter 'a can occur in a contravariant position, which means we need to derive a function like: cofmap :: (a -> b) -> (f b -> f a) This is pretty much the same as $fmap, only without the $(cofmap 'a 'a) case: $(cofmap 'a 'b) = \x -> x -- when b does not contain a $(cofmap 'a 'a) = error "type variable in contravariant position" $(cofmap 'a '(b1,b2)) = \x -> case x of (x1,x2) -> ($(cofmap 'a 'b1) x1, $(cofmap 'a 'b2) x2) $(cofmap 'a '[b]) = map $(cofmap 'a 'b) $(cofmap 'a '(T b1 b2)) = fmap $(cofmap 'a 'b2) -- when a only occurs in the last parameter, b2 $(cofmap 'a '(b -> c)) = \x b -> $(cofmap 'a' 'c) (x ($(fmap 'a 'c) b)) Note that the code produced by $(fmap _ _) is always a higher order function, with type `(a -> b) -> (g a -> g b)` for some g. When we need to do pattern matching on the type, this means create a lambda function (see the (,) case above). The resulting code for fmap can look a bit weird, for example: data X a = X (a,Int) -- generated instance instance Functor X where fmap f (X x) = (\y -> case y of (x1,x2) -> X (f x1, (\z -> z) x2)) x The optimizer should be able to simplify this code by simple inlining. An older version of the deriving code tried to avoid these applied lambda functions by producing a meta level function. But the function to be mapped, `f`, is a function on the code level, not on the meta level, so it was eta expanded to `\x -> [| f $x |]`. This resulted in too much eta expansion. It is better to produce too many lambdas than to eta expand, see ticket #7436. -} gen_Functor_binds :: SrcSpan -> TyCon -> (LHsBinds RdrName, BagDerivStuff) gen_Functor_binds loc tycon = (unitBag fmap_bind, emptyBag) where data_cons = tyConDataCons tycon fun_name = L loc fmap_RDR fmap_bind = mkRdrFunBind fun_name eqns fmap_eqn con = evalState (match_for_con [f_Pat] con =<< parts) bs_RDRs where parts = sequence $ foldDataConArgs ft_fmap con eqns | null data_cons = [mkSimpleMatch (FunRhs fun_name Prefix) [nlWildPat, nlWildPat] (error_Expr "Void fmap")] | otherwise = map fmap_eqn data_cons ft_fmap :: FFoldType (State [RdrName] (LHsExpr RdrName)) ft_fmap = FT { ft_triv = mkSimpleLam $ \x -> return x -- fmap f = \x -> x , ft_var = return f_Expr -- fmap f = f , ft_fun = \g h -> do gg <- g hh <- h mkSimpleLam2 $ \x b -> return $ nlHsApp hh (nlHsApp x (nlHsApp gg b)) -- fmap f = \x b -> h (x (g b)) , ft_tup = \t gs -> do gg <- sequence gs mkSimpleLam $ mkSimpleTupleCase match_for_con t gg -- fmap f = \x -> case x of (a1,a2,..) -> (g1 a1,g2 a2,..) , ft_ty_app = \_ g -> nlHsApp fmap_Expr <$> g -- fmap f = fmap g , ft_forall = \_ g -> g , ft_bad_app = panic "in other argument" , ft_co_var = panic "contravariant" } -- Con a1 a2 ... -> Con (f1 a1) (f2 a2) ... match_for_con :: [LPat RdrName] -> DataCon -> [LHsExpr RdrName] -> State [RdrName] (LMatch RdrName (LHsExpr RdrName)) match_for_con = mkSimpleConMatch CaseAlt $ \con_name xs -> return $ nlHsApps con_name xs -- Con x1 x2 .. {- Utility functions related to Functor deriving. Since several things use the same pattern of traversal, this is abstracted into functorLikeTraverse. This function works like a fold: it makes a value of type 'a' in a bottom up way. -} -- Generic traversal for Functor deriving -- See Note [FFoldType and functorLikeTraverse] data FFoldType a -- Describes how to fold over a Type in a functor like way = FT { ft_triv :: a -- ^ Does not contain variable , ft_var :: a -- ^ The variable itself , ft_co_var :: a -- ^ The variable itself, contravariantly , ft_fun :: a -> a -> a -- ^ Function type , ft_tup :: TyCon -> [a] -> a -- ^ Tuple type , ft_ty_app :: Type -> a -> a -- ^ Type app, variable only in last argument , ft_bad_app :: a -- ^ Type app, variable other than in last argument , ft_forall :: TcTyVar -> a -> a -- ^ Forall type } functorLikeTraverse :: forall a. TyVar -- ^ Variable to look for -> FFoldType a -- ^ How to fold -> Type -- ^ Type to process -> a functorLikeTraverse var (FT { ft_triv = caseTrivial, ft_var = caseVar , ft_co_var = caseCoVar, ft_fun = caseFun , ft_tup = caseTuple, ft_ty_app = caseTyApp , ft_bad_app = caseWrongArg, ft_forall = caseForAll }) ty = fst (go False ty) where go :: Bool -- Covariant or contravariant context -> Type -> (a, Bool) -- (result of type a, does type contain var) go co ty | Just ty' <- coreView ty = go co ty' go co (TyVarTy v) | v == var = (if co then caseCoVar else caseVar,True) go co (FunTy x y) | isPredTy x = go co y | xc || yc = (caseFun xr yr,True) where (xr,xc) = go (not co) x (yr,yc) = go co y go co (AppTy x y) | xc = (caseWrongArg, True) | yc = (caseTyApp x yr, True) where (_, xc) = go co x (yr,yc) = go co y go co ty@(TyConApp con args) | not (or xcs) = (caseTrivial, False) -- Variable does not occur -- At this point we know that xrs, xcs is not empty, -- and at least one xr is True | isTupleTyCon con = (caseTuple con xrs, True) | or (init xcs) = (caseWrongArg, True) -- T (..var..) ty | Just (fun_ty, _) <- splitAppTy_maybe ty -- T (..no var..) ty = (caseTyApp fun_ty (last xrs), True) | otherwise = (caseWrongArg, True) -- Non-decomposable (eg type function) where (xrs,xcs) = unzip (map (go co) args) go co (ForAllTy (TvBndr v vis) x) | isVisibleArgFlag vis = panic "unexpected visible binder" | v /= var && xc = (caseForAll v xr,True) where (xr,xc) = go co x go _ _ = (caseTrivial,False) -- Return all syntactic subterms of ty that contain var somewhere -- These are the things that should appear in instance constraints deepSubtypesContaining :: TyVar -> Type -> [TcType] deepSubtypesContaining tv = functorLikeTraverse tv (FT { ft_triv = [] , ft_var = [] , ft_fun = (++) , ft_tup = \_ xs -> concat xs , ft_ty_app = (:) , ft_bad_app = panic "in other argument" , ft_co_var = panic "contravariant" , ft_forall = \v xs -> filterOut ((v `elemVarSet`) . tyCoVarsOfType) xs }) foldDataConArgs :: FFoldType a -> DataCon -> [a] -- Fold over the arguments of the datacon foldDataConArgs ft con = map foldArg (dataConOrigArgTys con) where foldArg = case getTyVar_maybe (last (tyConAppArgs (dataConOrigResTy con))) of Just tv -> functorLikeTraverse tv ft Nothing -> const (ft_triv ft) -- If we are deriving Foldable for a GADT, there is a chance that the last -- type variable in the data type isn't actually a type variable at all. -- (for example, this can happen if the last type variable is refined to -- be a concrete type such as Int). If the last type variable is refined -- to be a specific type, then getTyVar_maybe will return Nothing. -- See Note [DeriveFoldable with ExistentialQuantification] -- -- The kind checks have ensured the last type parameter is of kind *. -- Make a HsLam using a fresh variable from a State monad mkSimpleLam :: (LHsExpr RdrName -> State [RdrName] (LHsExpr RdrName)) -> State [RdrName] (LHsExpr RdrName) -- (mkSimpleLam fn) returns (\x. fn(x)) mkSimpleLam lam = do (n:names) <- get put names body <- lam (nlHsVar n) return (mkHsLam [nlVarPat n] body) mkSimpleLam2 :: (LHsExpr RdrName -> LHsExpr RdrName -> State [RdrName] (LHsExpr RdrName)) -> State [RdrName] (LHsExpr RdrName) mkSimpleLam2 lam = do (n1:n2:names) <- get put names body <- lam (nlHsVar n1) (nlHsVar n2) return (mkHsLam [nlVarPat n1,nlVarPat n2] body) -- "Con a1 a2 a3 -> fold [x1 a1, x2 a2, x3 a3]" -- -- @mkSimpleConMatch fold extra_pats con insides@ produces a match clause in -- which the LHS pattern-matches on @extra_pats@, followed by a match on the -- constructor @con@ and its arguments. The RHS folds (with @fold@) over @con@ -- and its arguments, applying an expression (from @insides@) to each of the -- respective arguments of @con@. mkSimpleConMatch :: Monad m => HsMatchContext RdrName -> (RdrName -> [LHsExpr RdrName] -> m (LHsExpr RdrName)) -> [LPat RdrName] -> DataCon -> [LHsExpr RdrName] -> m (LMatch RdrName (LHsExpr RdrName)) mkSimpleConMatch ctxt fold extra_pats con insides = do let con_name = getRdrName con let vars_needed = takeList insides as_RDRs let pat = nlConVarPat con_name vars_needed rhs <- fold con_name (zipWith nlHsApp insides (map nlHsVar vars_needed)) return $ mkMatch ctxt (extra_pats ++ [pat]) rhs (noLoc emptyLocalBinds) -- "Con a1 a2 a3 -> fmap (\b2 -> Con a1 b2 a3) (traverse f a2)" -- -- @mkSimpleConMatch2 fold extra_pats con insides@ behaves very similarly to -- 'mkSimpleConMatch', with two key differences: -- -- 1. @insides@ is a @[Maybe (LHsExpr RdrName)]@ instead of a -- @[LHsExpr RdrName]@. This is because it filters out the expressions -- corresponding to arguments whose types do not mention the last type -- variable in a derived 'Foldable' or 'Traversable' instance (i.e., the -- 'Nothing' elements of @insides@). -- -- 2. @fold@ takes an expression as its first argument instead of a -- constructor name. This is because it uses a specialized -- constructor function expression that only takes as many parameters as -- there are argument types that mention the last type variable. -- -- See Note [Generated code for DeriveFoldable and DeriveTraversable] mkSimpleConMatch2 :: Monad m => HsMatchContext RdrName -> (LHsExpr RdrName -> [LHsExpr RdrName] -> m (LHsExpr RdrName)) -> [LPat RdrName] -> DataCon -> [Maybe (LHsExpr RdrName)] -> m (LMatch RdrName (LHsExpr RdrName)) mkSimpleConMatch2 ctxt fold extra_pats con insides = do let con_name = getRdrName con vars_needed = takeList insides as_RDRs pat = nlConVarPat con_name vars_needed -- Make sure to zip BEFORE invoking catMaybes. We want the variable -- indicies in each expression to match up with the argument indices -- in con_expr (defined below). exps = catMaybes $ zipWith (\i v -> (`nlHsApp` v) <$> i) insides (map nlHsVar vars_needed) -- An element of argTysTyVarInfo is True if the constructor argument -- with the same index has a type which mentions the last type -- variable. argTysTyVarInfo = map isJust insides (asWithTyVar, asWithoutTyVar) = partitionByList argTysTyVarInfo as_RDRs con_expr | null asWithTyVar = nlHsApps con_name $ map nlHsVar asWithoutTyVar | otherwise = let bs = filterByList argTysTyVarInfo bs_RDRs vars = filterByLists argTysTyVarInfo (map nlHsVar bs_RDRs) (map nlHsVar as_RDRs) in mkHsLam (map nlVarPat bs) (nlHsApps con_name vars) rhs <- fold con_expr exps return $ mkMatch ctxt (extra_pats ++ [pat]) rhs (noLoc emptyLocalBinds) -- "case x of (a1,a2,a3) -> fold [x1 a1, x2 a2, x3 a3]" mkSimpleTupleCase :: Monad m => ([LPat RdrName] -> DataCon -> [a] -> m (LMatch RdrName (LHsExpr RdrName))) -> TyCon -> [a] -> LHsExpr RdrName -> m (LHsExpr RdrName) mkSimpleTupleCase match_for_con tc insides x = do { let data_con = tyConSingleDataCon tc ; match <- match_for_con [] data_con insides ; return $ nlHsCase x [match] } {- ************************************************************************ * * Foldable instances see http://www.mail-archive.com/haskell-prime@haskell.org/msg02116.html * * ************************************************************************ Deriving Foldable instances works the same way as Functor instances, only Foldable instances are not possible for function types at all. Given (data T a = T a a (T a) deriving Foldable), we get: instance Foldable T where foldr f z (T x1 x2 x3) = $(foldr 'a 'a) x1 ( $(foldr 'a 'a) x2 ( $(foldr 'a '(T a)) x3 z ) ) -XDeriveFoldable is different from -XDeriveFunctor in that it filters out arguments to the constructor that would produce useless code in a Foldable instance. For example, the following datatype: data Foo a = Foo Int a Int deriving Foldable would have the following generated Foldable instance: instance Foldable Foo where foldr f z (Foo x1 x2 x3) = $(foldr 'a 'a) x2 since neither of the two Int arguments are folded over. The cases are: $(foldr 'a 'a) = f $(foldr 'a '(b1,b2)) = \x z -> case x of (x1,x2) -> $(foldr 'a 'b1) x1 ( $(foldr 'a 'b2) x2 z ) $(foldr 'a '(T b1 b2)) = \x z -> foldr $(foldr 'a 'b2) z x -- when a only occurs in the last parameter, b2 Note that the arguments to the real foldr function are the wrong way around, since (f :: a -> b -> b), while (foldr f :: b -> t a -> b). One can envision a case for types that don't contain the last type variable: $(foldr 'a 'b) = \x z -> z -- when b does not contain a But this case will never materialize, since the aforementioned filtering removes all such types from consideration. See Note [Generated code for DeriveFoldable and DeriveTraversable]. Foldable instances differ from Functor and Traversable instances in that Foldable instances can be derived for data types in which the last type variable is existentially quantified. In particular, if the last type variable is refined to a more specific type in a GADT: data GADT a where G :: a ~ Int => a -> G Int then the deriving machinery does not attempt to check that the type a contains Int, since it is not syntactically equal to a type variable. That is, the derived Foldable instance for GADT is: instance Foldable GADT where foldr _ z (GADT _) = z See Note [DeriveFoldable with ExistentialQuantification]. -} gen_Foldable_binds :: SrcSpan -> TyCon -> (LHsBinds RdrName, BagDerivStuff) gen_Foldable_binds loc tycon = (listToBag [foldr_bind, foldMap_bind], emptyBag) where data_cons = tyConDataCons tycon foldr_bind = mkRdrFunBind (L loc foldable_foldr_RDR) eqns eqns = map foldr_eqn data_cons foldr_eqn con = evalState (match_foldr z_Expr [f_Pat,z_Pat] con =<< parts) bs_RDRs where parts = sequence $ foldDataConArgs ft_foldr con foldMap_bind = mkRdrFunBind (L loc foldMap_RDR) (map foldMap_eqn data_cons) foldMap_eqn con = evalState (match_foldMap [f_Pat] con =<< parts) bs_RDRs where parts = sequence $ foldDataConArgs ft_foldMap con -- Yields 'Just' an expression if we're folding over a type that mentions -- the last type parameter of the datatype. Otherwise, yields 'Nothing'. -- See Note [FFoldType and functorLikeTraverse] ft_foldr :: FFoldType (State [RdrName] (Maybe (LHsExpr RdrName))) ft_foldr = FT { ft_triv = return Nothing -- foldr f = \x z -> z , ft_var = return $ Just f_Expr -- foldr f = f , ft_tup = \t g -> do gg <- sequence g lam <- mkSimpleLam2 $ \x z -> mkSimpleTupleCase (match_foldr z) t gg x return (Just lam) -- foldr f = (\x z -> case x of ...) , ft_ty_app = \_ g -> do gg <- g mapM (\gg' -> mkSimpleLam2 $ \x z -> return $ nlHsApps foldable_foldr_RDR [gg',z,x]) gg -- foldr f = (\x z -> foldr g z x) , ft_forall = \_ g -> g , ft_co_var = panic "contravariant" , ft_fun = panic "function" , ft_bad_app = panic "in other argument" } match_foldr :: LHsExpr RdrName -> [LPat RdrName] -> DataCon -> [Maybe (LHsExpr RdrName)] -> State [RdrName] (LMatch RdrName (LHsExpr RdrName)) match_foldr z = mkSimpleConMatch2 LambdaExpr $ \_ xs -> return (mkFoldr xs) where -- g1 v1 (g2 v2 (.. z)) mkFoldr :: [LHsExpr RdrName] -> LHsExpr RdrName mkFoldr = foldr nlHsApp z -- See Note [FFoldType and functorLikeTraverse] ft_foldMap :: FFoldType (State [RdrName] (Maybe (LHsExpr RdrName))) ft_foldMap = FT { ft_triv = return Nothing -- foldMap f = \x -> mempty , ft_var = return (Just f_Expr) -- foldMap f = f , ft_tup = \t g -> do gg <- sequence g lam <- mkSimpleLam $ mkSimpleTupleCase match_foldMap t gg return (Just lam) -- foldMap f = \x -> case x of (..,) , ft_ty_app = \_ g -> fmap (nlHsApp foldMap_Expr) <$> g -- foldMap f = foldMap g , ft_forall = \_ g -> g , ft_co_var = panic "contravariant" , ft_fun = panic "function" , ft_bad_app = panic "in other argument" } match_foldMap :: [LPat RdrName] -> DataCon -> [Maybe (LHsExpr RdrName)] -> State [RdrName] (LMatch RdrName (LHsExpr RdrName)) match_foldMap = mkSimpleConMatch2 CaseAlt $ \_ xs -> return (mkFoldMap xs) where -- mappend v1 (mappend v2 ..) mkFoldMap :: [LHsExpr RdrName] -> LHsExpr RdrName mkFoldMap [] = mempty_Expr mkFoldMap xs = foldr1 (\x y -> nlHsApps mappend_RDR [x,y]) xs {- ************************************************************************ * * Traversable instances see http://www.mail-archive.com/haskell-prime@haskell.org/msg02116.html * * ************************************************************************ Again, Traversable is much like Functor and Foldable. The cases are: $(traverse 'a 'a) = f $(traverse 'a '(b1,b2)) = \x -> case x of (x1,x2) -> (,) <$> $(traverse 'a 'b1) x1 <*> $(traverse 'a 'b2) x2 $(traverse 'a '(T b1 b2)) = traverse $(traverse 'a 'b2) -- when a only occurs in the last parameter, b2 Like -XDeriveFoldable, -XDeriveTraversable filters out arguments whose types do not mention the last type parameter. Therefore, the following datatype: data Foo a = Foo Int a Int would have the following derived Traversable instance: instance Traversable Foo where traverse f (Foo x1 x2 x3) = fmap (\b2 -> Foo x1 b2 x3) ( $(traverse 'a 'a) x2 ) since the two Int arguments do not produce any effects in a traversal. One can envision a case for types that do not mention the last type parameter: $(traverse 'a 'b) = pure -- when b does not contain a But this case will never materialize, since the aforementioned filtering removes all such types from consideration. See Note [Generated code for DeriveFoldable and DeriveTraversable]. -} gen_Traversable_binds :: SrcSpan -> TyCon -> (LHsBinds RdrName, BagDerivStuff) gen_Traversable_binds loc tycon = (unitBag traverse_bind, emptyBag) where data_cons = tyConDataCons tycon traverse_bind = mkRdrFunBind (L loc traverse_RDR) eqns eqns = map traverse_eqn data_cons traverse_eqn con = evalState (match_for_con [f_Pat] con =<< parts) bs_RDRs where parts = sequence $ foldDataConArgs ft_trav con -- Yields 'Just' an expression if we're folding over a type that mentions -- the last type parameter of the datatype. Otherwise, yields 'Nothing'. -- See Note [FFoldType and functorLikeTraverse] ft_trav :: FFoldType (State [RdrName] (Maybe (LHsExpr RdrName))) ft_trav = FT { ft_triv = return Nothing -- traverse f = pure x , ft_var = return (Just f_Expr) -- traverse f = f x , ft_tup = \t gs -> do gg <- sequence gs lam <- mkSimpleLam $ mkSimpleTupleCase match_for_con t gg return (Just lam) -- traverse f = \x -> case x of (a1,a2,..) -> -- (,,) <$> g1 a1 <*> g2 a2 <*> .. , ft_ty_app = \_ g -> fmap (nlHsApp traverse_Expr) <$> g -- traverse f = traverse g , ft_forall = \_ g -> g , ft_co_var = panic "contravariant" , ft_fun = panic "function" , ft_bad_app = panic "in other argument" } -- Con a1 a2 ... -> fmap (\b1 b2 ... -> Con b1 b2 ...) (g1 a1) -- <*> g2 a2 <*> ... match_for_con :: [LPat RdrName] -> DataCon -> [Maybe (LHsExpr RdrName)] -> State [RdrName] (LMatch RdrName (LHsExpr RdrName)) match_for_con = mkSimpleConMatch2 CaseAlt $ \con xs -> return (mkApCon con xs) where -- fmap (\b1 b2 ... -> Con b1 b2 ...) x1 <*> x2 <*> .. mkApCon :: LHsExpr RdrName -> [LHsExpr RdrName] -> LHsExpr RdrName mkApCon con [] = nlHsApps pure_RDR [con] mkApCon con (x:xs) = foldl appAp (nlHsApps fmap_RDR [con,x]) xs where appAp x y = nlHsApps ap_RDR [x,y] {- ************************************************************************ * * Lift instances * * ************************************************************************ Example: data Foo a = Foo a | a :^: a deriving Lift ==> instance (Lift a) => Lift (Foo a) where lift (Foo a) = appE (conE (mkNameG_d "package-name" "ModuleName" "Foo")) (lift a) lift (u :^: v) = infixApp (lift u) (conE (mkNameG_d "package-name" "ModuleName" ":^:")) (lift v) Note that (mkNameG_d "package-name" "ModuleName" "Foo") is equivalent to what 'Foo would be when using the -XTemplateHaskell extension. To make sure that -XDeriveLift can be used on stage-1 compilers, however, we expliticly invoke makeG_d. -} gen_Lift_binds :: SrcSpan -> TyCon -> (LHsBinds RdrName, BagDerivStuff) gen_Lift_binds loc tycon | null data_cons = (unitBag (L loc $ mkFunBind (L loc lift_RDR) [mkMatch (FunRhs (L loc lift_RDR) Prefix) [nlWildPat] errorMsg_Expr (noLoc emptyLocalBinds)]) , emptyBag) | otherwise = (unitBag lift_bind, emptyBag) where errorMsg_Expr = nlHsVar error_RDR `nlHsApp` nlHsLit (mkHsString $ "Can't lift value of empty datatype " ++ tycon_str) lift_bind = mk_FunBind loc lift_RDR (map pats_etc data_cons) data_cons = tyConDataCons tycon tycon_str = occNameString . nameOccName . tyConName $ tycon pats_etc data_con = ([con_pat], lift_Expr) where con_pat = nlConVarPat data_con_RDR as_needed data_con_RDR = getRdrName data_con con_arity = dataConSourceArity data_con as_needed = take con_arity as_RDRs lifted_as = zipWithEqual "mk_lift_app" mk_lift_app tys_needed as_needed tycon_name = tyConName tycon is_infix = dataConIsInfix data_con tys_needed = dataConOrigArgTys data_con mk_lift_app ty a | not (isUnliftedType ty) = nlHsApp (nlHsVar lift_RDR) (nlHsVar a) | otherwise = nlHsApp (nlHsVar litE_RDR) (primLitOp (mkBoxExp (nlHsVar a))) where (primLitOp, mkBoxExp) = primLitOps "Lift" tycon ty pkg_name = unitIdString . moduleUnitId . nameModule $ tycon_name mod_name = moduleNameString . moduleName . nameModule $ tycon_name con_name = occNameString . nameOccName . dataConName $ data_con conE_Expr = nlHsApp (nlHsVar conE_RDR) (nlHsApps mkNameG_dRDR (map (nlHsLit . mkHsString) [pkg_name, mod_name, con_name])) lift_Expr | is_infix = nlHsApps infixApp_RDR [a1, conE_Expr, a2] | otherwise = foldl mk_appE_app conE_Expr lifted_as (a1:a2:_) = lifted_as mk_appE_app :: LHsExpr RdrName -> LHsExpr RdrName -> LHsExpr RdrName mk_appE_app a b = nlHsApps appE_RDR [a, b] {- ************************************************************************ * * Newtype-deriving instances * * ************************************************************************ Note [Newtype-deriving instances] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ We take every method in the original instance and `coerce` it to fit into the derived instance. We need a type annotation on the argument to `coerce` to make it obvious what instantiation of the method we're coercing from. So from, say, class C a b where op :: a -> [b] -> Int newtype T x = MkT <rep-ty> instance C a <rep-ty> => C a (T x) where op = (coerce (op :: a -> [<rep-ty>] -> Int) ) :: a -> [T x] -> Int Notice that we give the 'coerce' call two type signatures: one to fix the of the inner call, and one for the expected type. The outer type signature ought to be redundant, but may improve error messages. The inner one is essential to fix the type at which 'op' is called. See #8503 for more discussion. Here's a wrinkle. Supppose 'op' is locally overloaded: class C2 b where op2 :: forall a. Eq a => a -> [b] -> Int Then we could do exactly as above, but it's a bit redundant to instantiate op, then re-generalise with the inner signature. (The inner sig is only there to fix the type at which 'op' is called.) So we just instantiate the signature, and add instance C2 <rep-ty> => C2 (T x) where op2 = (coerce (op2 :: a -> [<rep-ty>] -> Int) ) :: forall a. Eq a => a -> [T x] -> Int -} gen_Newtype_binds :: SrcSpan -> Class -- the class being derived -> [TyVar] -- the tvs in the instance head -> [Type] -- instance head parameters (incl. newtype) -> Type -- the representation type (already eta-reduced) -> LHsBinds RdrName -- See Note [Newtype-deriving instances] gen_Newtype_binds loc cls inst_tvs cls_tys rhs_ty = listToBag $ map mk_bind (classMethods cls) where coerce_RDR = getRdrName coerceId mk_bind :: Id -> LHsBind RdrName mk_bind meth_id = mkRdrFunBind (L loc meth_RDR) [mkSimpleMatch (FunRhs (L loc meth_RDR) Prefix) [] rhs_expr] where Pair from_ty to_ty = mkCoerceClassMethEqn cls inst_tvs cls_tys rhs_ty meth_id -- See "wrinkle" in Note [Newtype-deriving instances] (_, _, from_ty') = tcSplitSigmaTy from_ty meth_RDR = getRdrName meth_id rhs_expr = ( nlHsVar coerce_RDR `nlHsApp` (nlHsVar meth_RDR `nlExprWithTySig` toLHsSigWcType from_ty')) `nlExprWithTySig` toLHsSigWcType to_ty nlExprWithTySig :: LHsExpr RdrName -> LHsSigWcType RdrName -> LHsExpr RdrName nlExprWithTySig e s = noLoc (ExprWithTySig e s) mkCoerceClassMethEqn :: Class -- the class being derived -> [TyVar] -- the tvs in the instance head -> [Type] -- instance head parameters (incl. newtype) -> Type -- the representation type (already eta-reduced) -> Id -- the method to look at -> Pair Type -- See Note [Newtype-deriving instances] -- The pair is the (from_type, to_type), where to_type is -- the type of the method we are tyrying to get mkCoerceClassMethEqn cls inst_tvs cls_tys rhs_ty id = Pair (substTy rhs_subst user_meth_ty) (substTy lhs_subst user_meth_ty) where cls_tvs = classTyVars cls in_scope = mkInScopeSet $ mkVarSet inst_tvs lhs_subst = mkTvSubst in_scope (zipTyEnv cls_tvs cls_tys) rhs_subst = mkTvSubst in_scope (zipTyEnv cls_tvs (changeLast cls_tys rhs_ty)) (_class_tvs, _class_constraint, user_meth_ty) = tcSplitMethodTy (varType id) changeLast :: [a] -> a -> [a] changeLast [] _ = panic "changeLast" changeLast [_] x = [x] changeLast (x:xs) x' = x : changeLast xs x' {- ************************************************************************ * * \subsection{Generating extra binds (@con2tag@ and @tag2con@)} * * ************************************************************************ \begin{verbatim} data Foo ... = ... con2tag_Foo :: Foo ... -> Int# tag2con_Foo :: Int -> Foo ... -- easier if Int, not Int# maxtag_Foo :: Int -- ditto (NB: not unlifted) \end{verbatim} The `tags' here start at zero, hence the @fIRST_TAG@ (currently one) fiddling around. -} genAuxBindSpec :: SrcSpan -> AuxBindSpec -> (LHsBind RdrName, LSig RdrName) genAuxBindSpec loc (DerivCon2Tag tycon) = (mk_FunBind loc rdr_name eqns, L loc (TypeSig [L loc rdr_name] sig_ty)) where rdr_name = con2tag_RDR tycon sig_ty = mkLHsSigWcType $ L loc $ HsCoreTy $ mkSpecSigmaTy (tyConTyVars tycon) (tyConStupidTheta tycon) $ mkParentType tycon `mkFunTy` intPrimTy lots_of_constructors = tyConFamilySize tycon > 8 -- was: mAX_FAMILY_SIZE_FOR_VEC_RETURNS -- but we don't do vectored returns any more. eqns | lots_of_constructors = [get_tag_eqn] | otherwise = map mk_eqn (tyConDataCons tycon) get_tag_eqn = ([nlVarPat a_RDR], nlHsApp (nlHsVar getTag_RDR) a_Expr) mk_eqn :: DataCon -> ([LPat RdrName], LHsExpr RdrName) mk_eqn con = ([nlWildConPat con], nlHsLit (HsIntPrim "" (toInteger ((dataConTag con) - fIRST_TAG)))) genAuxBindSpec loc (DerivTag2Con tycon) = (mk_FunBind loc rdr_name [([nlConVarPat intDataCon_RDR [a_RDR]], nlHsApp (nlHsVar tagToEnum_RDR) a_Expr)], L loc (TypeSig [L loc rdr_name] sig_ty)) where sig_ty = mkLHsSigWcType $ L loc $ HsCoreTy $ mkSpecForAllTys (tyConTyVars tycon) $ intTy `mkFunTy` mkParentType tycon rdr_name = tag2con_RDR tycon genAuxBindSpec loc (DerivMaxTag tycon) = (mkHsVarBind loc rdr_name rhs, L loc (TypeSig [L loc rdr_name] sig_ty)) where rdr_name = maxtag_RDR tycon sig_ty = mkLHsSigWcType (L loc (HsCoreTy intTy)) rhs = nlHsApp (nlHsVar intDataCon_RDR) (nlHsLit (HsIntPrim "" max_tag)) max_tag = case (tyConDataCons tycon) of data_cons -> toInteger ((length data_cons) - fIRST_TAG) type SeparateBagsDerivStuff = -- AuxBinds and SYB bindings ( Bag (LHsBind RdrName, LSig RdrName) -- Extra bindings (used by Generic only) , Bag (FamInst) -- Extra family instances , Bag (InstInfo RdrName)) -- Extra instances genAuxBinds :: SrcSpan -> BagDerivStuff -> SeparateBagsDerivStuff genAuxBinds loc b = genAuxBinds' b2 where (b1,b2) = partitionBagWith splitDerivAuxBind b splitDerivAuxBind (DerivAuxBind x) = Left x splitDerivAuxBind x = Right x rm_dups = foldrBag dup_check emptyBag dup_check a b = if anyBag (== a) b then b else consBag a b genAuxBinds' :: BagDerivStuff -> SeparateBagsDerivStuff genAuxBinds' = foldrBag f ( mapBag (genAuxBindSpec loc) (rm_dups b1) , emptyBag, emptyBag) f :: DerivStuff -> SeparateBagsDerivStuff -> SeparateBagsDerivStuff f (DerivAuxBind _) = panic "genAuxBinds'" -- We have removed these before f (DerivHsBind b) = add1 b f (DerivFamInst t) = add2 t f (DerivInst i) = add3 i add1 x (a,b,c) = (x `consBag` a,b,c) add2 x (a,b,c) = (a,x `consBag` b,c) add3 x (a,b,c) = (a,b,x `consBag` c) mkParentType :: TyCon -> Type -- Turn the representation tycon of a family into -- a use of its family constructor mkParentType tc = case tyConFamInst_maybe tc of Nothing -> mkTyConApp tc (mkTyVarTys (tyConTyVars tc)) Just (fam_tc,tys) -> mkTyConApp fam_tc tys {- ************************************************************************ * * \subsection{Utility bits for generating bindings} * * ************************************************************************ -} mk_FunBind :: SrcSpan -> RdrName -> [([LPat RdrName], LHsExpr RdrName)] -> LHsBind RdrName mk_FunBind = mk_HRFunBind 0 -- by using mk_FunBind and not mk_HRFunBind, -- the caller says that the Void case needs no -- patterns -- | This variant of 'mk_FunBind' puts an 'Arity' number of wildcards before -- the "=" in the empty-data-decl case. This is necessary if the function -- has a higher-rank type, like foldl. (See deriving/should_compile/T4302) mk_HRFunBind :: Arity -> SrcSpan -> RdrName -> [([LPat RdrName], LHsExpr RdrName)] -> LHsBind RdrName mk_HRFunBind arity loc fun pats_and_exprs = mkHRRdrFunBind arity (L loc fun) matches where matches = [mkMatch (FunRhs (L loc fun) Prefix) p e (noLoc emptyLocalBinds) | (p,e) <-pats_and_exprs] mkRdrFunBind :: Located RdrName -> [LMatch RdrName (LHsExpr RdrName)] -> LHsBind RdrName mkRdrFunBind = mkHRRdrFunBind 0 mkHRRdrFunBind :: Arity -> Located RdrName -> [LMatch RdrName (LHsExpr RdrName)] -> LHsBind RdrName mkHRRdrFunBind arity fun@(L loc fun_rdr) matches = L loc (mkFunBind fun matches') where -- Catch-all eqn looks like -- fmap = error "Void fmap" -- It's needed if there no data cons at all, -- which can happen with -XEmptyDataDecls -- See Trac #4302 matches' = if null matches then [mkMatch (FunRhs fun Prefix) (replicate arity nlWildPat) (error_Expr str) (noLoc emptyLocalBinds)] else matches str = "Void " ++ occNameString (rdrNameOcc fun_rdr) box :: String -- The class involved -> TyCon -- The tycon involved -> LHsExpr RdrName -- The argument -> Type -- The argument type -> LHsExpr RdrName -- Boxed version of the arg -- See Note [Deriving and unboxed types] in TcDeriv box cls_str tycon arg arg_ty = nlHsApp (nlHsVar box_con) arg where box_con = assoc_ty_id cls_str tycon boxConTbl arg_ty --------------------- primOrdOps :: String -- The class involved -> TyCon -- The tycon involved -> Type -- The type -> (RdrName, RdrName, RdrName, RdrName, RdrName) -- (lt,le,eq,ge,gt) -- See Note [Deriving and unboxed types] in TcDeriv primOrdOps str tycon ty = assoc_ty_id str tycon ordOpTbl ty primLitOps :: String -- The class involved -> TyCon -- The tycon involved -> Type -- The type -> ( LHsExpr RdrName -> LHsExpr RdrName -- Constructs a Q Exp value , LHsExpr RdrName -> LHsExpr RdrName -- Constructs a boxed value ) primLitOps str tycon ty = ( assoc_ty_id str tycon litConTbl ty , \v -> nlHsVar boxRDR `nlHsApp` v ) where boxRDR | ty `eqType` addrPrimTy = unpackCString_RDR | otherwise = assoc_ty_id str tycon boxConTbl ty ordOpTbl :: [(Type, (RdrName, RdrName, RdrName, RdrName, RdrName))] ordOpTbl = [(charPrimTy , (ltChar_RDR , leChar_RDR , eqChar_RDR , geChar_RDR , gtChar_RDR )) ,(intPrimTy , (ltInt_RDR , leInt_RDR , eqInt_RDR , geInt_RDR , gtInt_RDR )) ,(wordPrimTy , (ltWord_RDR , leWord_RDR , eqWord_RDR , geWord_RDR , gtWord_RDR )) ,(addrPrimTy , (ltAddr_RDR , leAddr_RDR , eqAddr_RDR , geAddr_RDR , gtAddr_RDR )) ,(floatPrimTy , (ltFloat_RDR , leFloat_RDR , eqFloat_RDR , geFloat_RDR , gtFloat_RDR )) ,(doublePrimTy, (ltDouble_RDR, leDouble_RDR, eqDouble_RDR, geDouble_RDR, gtDouble_RDR)) ] boxConTbl :: [(Type, RdrName)] boxConTbl = [(charPrimTy , getRdrName charDataCon ) ,(intPrimTy , getRdrName intDataCon ) ,(wordPrimTy , getRdrName wordDataCon ) ,(floatPrimTy , getRdrName floatDataCon ) ,(doublePrimTy, getRdrName doubleDataCon) ] -- | A table of postfix modifiers for unboxed values. postfixModTbl :: [(Type, String)] postfixModTbl = [(charPrimTy , "#" ) ,(intPrimTy , "#" ) ,(wordPrimTy , "##") ,(floatPrimTy , "#" ) ,(doublePrimTy, "##") ] litConTbl :: [(Type, LHsExpr RdrName -> LHsExpr RdrName)] litConTbl = [(charPrimTy , nlHsApp (nlHsVar charPrimL_RDR)) ,(intPrimTy , nlHsApp (nlHsVar intPrimL_RDR) . nlHsApp (nlHsVar toInteger_RDR)) ,(wordPrimTy , nlHsApp (nlHsVar wordPrimL_RDR) . nlHsApp (nlHsVar toInteger_RDR)) ,(addrPrimTy , nlHsApp (nlHsVar stringPrimL_RDR) . nlHsApp (nlHsApp (nlHsVar map_RDR) (compose_RDR `nlHsApps` [ nlHsVar fromIntegral_RDR , nlHsVar fromEnum_RDR ]))) ,(floatPrimTy , nlHsApp (nlHsVar floatPrimL_RDR) . nlHsApp (nlHsVar toRational_RDR)) ,(doublePrimTy, nlHsApp (nlHsVar doublePrimL_RDR) . nlHsApp (nlHsVar toRational_RDR)) ] -- | Lookup `Type` in an association list. assoc_ty_id :: String -- The class involved -> TyCon -- The tycon involved -> [(Type,a)] -- The table -> Type -- The type -> a -- The result of the lookup assoc_ty_id cls_str _ tbl ty | null res = pprPanic "Error in deriving:" (text "Can't derive" <+> text cls_str <+> text "for primitive type" <+> ppr ty) | otherwise = head res where res = [id | (ty',id) <- tbl, ty `eqType` ty'] ----------------------------------------------------------------------- and_Expr :: LHsExpr RdrName -> LHsExpr RdrName -> LHsExpr RdrName and_Expr a b = genOpApp a and_RDR b ----------------------------------------------------------------------- eq_Expr :: TyCon -> Type -> LHsExpr RdrName -> LHsExpr RdrName -> LHsExpr RdrName eq_Expr tycon ty a b | not (isUnliftedType ty) = genOpApp a eq_RDR b | otherwise = genPrimOpApp a prim_eq b where (_, _, prim_eq, _, _) = primOrdOps "Eq" tycon ty untag_Expr :: TyCon -> [( RdrName, RdrName)] -> LHsExpr RdrName -> LHsExpr RdrName untag_Expr _ [] expr = expr untag_Expr tycon ((untag_this, put_tag_here) : more) expr = nlHsCase (nlHsPar (nlHsVarApps (con2tag_RDR tycon) [untag_this])) {-of-} [mkHsCaseAlt (nlVarPat put_tag_here) (untag_Expr tycon more expr)] enum_from_to_Expr :: LHsExpr RdrName -> LHsExpr RdrName -> LHsExpr RdrName enum_from_then_to_Expr :: LHsExpr RdrName -> LHsExpr RdrName -> LHsExpr RdrName -> LHsExpr RdrName enum_from_to_Expr f t2 = nlHsApp (nlHsApp (nlHsVar enumFromTo_RDR) f) t2 enum_from_then_to_Expr f t t2 = nlHsApp (nlHsApp (nlHsApp (nlHsVar enumFromThenTo_RDR) f) t) t2 showParen_Expr :: LHsExpr RdrName -> LHsExpr RdrName -> LHsExpr RdrName showParen_Expr e1 e2 = nlHsApp (nlHsApp (nlHsVar showParen_RDR) e1) e2 nested_compose_Expr :: [LHsExpr RdrName] -> LHsExpr RdrName nested_compose_Expr [] = panic "nested_compose_expr" -- Arg is always non-empty nested_compose_Expr [e] = parenify e nested_compose_Expr (e:es) = nlHsApp (nlHsApp (nlHsVar compose_RDR) (parenify e)) (nested_compose_Expr es) -- impossible_Expr is used in case RHSs that should never happen. -- We generate these to keep the desugarer from complaining that they *might* happen! error_Expr :: String -> LHsExpr RdrName error_Expr string = nlHsApp (nlHsVar error_RDR) (nlHsLit (mkHsString string)) -- illegal_Expr is used when signalling error conditions in the RHS of a derived -- method. It is currently only used by Enum.{succ,pred} illegal_Expr :: String -> String -> String -> LHsExpr RdrName illegal_Expr meth tp msg = nlHsApp (nlHsVar error_RDR) (nlHsLit (mkHsString (meth ++ '{':tp ++ "}: " ++ msg))) -- illegal_toEnum_tag is an extended version of illegal_Expr, which also allows you -- to include the value of a_RDR in the error string. illegal_toEnum_tag :: String -> RdrName -> LHsExpr RdrName illegal_toEnum_tag tp maxtag = nlHsApp (nlHsVar error_RDR) (nlHsApp (nlHsApp (nlHsVar append_RDR) (nlHsLit (mkHsString ("toEnum{" ++ tp ++ "}: tag (")))) (nlHsApp (nlHsApp (nlHsApp (nlHsVar showsPrec_RDR) (nlHsIntLit 0)) (nlHsVar a_RDR)) (nlHsApp (nlHsApp (nlHsVar append_RDR) (nlHsLit (mkHsString ") is outside of enumeration's range (0,"))) (nlHsApp (nlHsApp (nlHsApp (nlHsVar showsPrec_RDR) (nlHsIntLit 0)) (nlHsVar maxtag)) (nlHsLit (mkHsString ")")))))) parenify :: LHsExpr RdrName -> LHsExpr RdrName parenify e@(L _ (HsVar _)) = e parenify e = mkHsPar e -- genOpApp wraps brackets round the operator application, so that the -- renamer won't subsequently try to re-associate it. genOpApp :: LHsExpr RdrName -> RdrName -> LHsExpr RdrName -> LHsExpr RdrName genOpApp e1 op e2 = nlHsPar (nlHsOpApp e1 op e2) genPrimOpApp :: LHsExpr RdrName -> RdrName -> LHsExpr RdrName -> LHsExpr RdrName genPrimOpApp e1 op e2 = nlHsPar (nlHsApp (nlHsVar tagToEnum_RDR) (nlHsOpApp e1 op e2)) a_RDR, b_RDR, c_RDR, d_RDR, f_RDR, k_RDR, z_RDR, ah_RDR, bh_RDR, ch_RDR, dh_RDR :: RdrName a_RDR = mkVarUnqual (fsLit "a") b_RDR = mkVarUnqual (fsLit "b") c_RDR = mkVarUnqual (fsLit "c") d_RDR = mkVarUnqual (fsLit "d") f_RDR = mkVarUnqual (fsLit "f") k_RDR = mkVarUnqual (fsLit "k") z_RDR = mkVarUnqual (fsLit "z") ah_RDR = mkVarUnqual (fsLit "a#") bh_RDR = mkVarUnqual (fsLit "b#") ch_RDR = mkVarUnqual (fsLit "c#") dh_RDR = mkVarUnqual (fsLit "d#") as_RDRs, bs_RDRs, cs_RDRs :: [RdrName] as_RDRs = [ mkVarUnqual (mkFastString ("a"++show i)) | i <- [(1::Int) .. ] ] bs_RDRs = [ mkVarUnqual (mkFastString ("b"++show i)) | i <- [(1::Int) .. ] ] cs_RDRs = [ mkVarUnqual (mkFastString ("c"++show i)) | i <- [(1::Int) .. ] ] a_Expr, c_Expr, f_Expr, z_Expr, ltTag_Expr, eqTag_Expr, gtTag_Expr, false_Expr, true_Expr, fmap_Expr, mempty_Expr, foldMap_Expr, traverse_Expr :: LHsExpr RdrName a_Expr = nlHsVar a_RDR -- b_Expr = nlHsVar b_RDR c_Expr = nlHsVar c_RDR f_Expr = nlHsVar f_RDR z_Expr = nlHsVar z_RDR ltTag_Expr = nlHsVar ltTag_RDR eqTag_Expr = nlHsVar eqTag_RDR gtTag_Expr = nlHsVar gtTag_RDR false_Expr = nlHsVar false_RDR true_Expr = nlHsVar true_RDR fmap_Expr = nlHsVar fmap_RDR -- pure_Expr = nlHsVar pure_RDR mempty_Expr = nlHsVar mempty_RDR foldMap_Expr = nlHsVar foldMap_RDR traverse_Expr = nlHsVar traverse_RDR a_Pat, b_Pat, c_Pat, d_Pat, f_Pat, k_Pat, z_Pat :: LPat RdrName a_Pat = nlVarPat a_RDR b_Pat = nlVarPat b_RDR c_Pat = nlVarPat c_RDR d_Pat = nlVarPat d_RDR f_Pat = nlVarPat f_RDR k_Pat = nlVarPat k_RDR z_Pat = nlVarPat z_RDR minusInt_RDR, tagToEnum_RDR :: RdrName minusInt_RDR = getRdrName (primOpId IntSubOp ) tagToEnum_RDR = getRdrName (primOpId TagToEnumOp) con2tag_RDR, tag2con_RDR, maxtag_RDR :: TyCon -> RdrName -- Generates Orig s RdrName, for the binding positions con2tag_RDR tycon = mk_tc_deriv_name tycon mkCon2TagOcc tag2con_RDR tycon = mk_tc_deriv_name tycon mkTag2ConOcc maxtag_RDR tycon = mk_tc_deriv_name tycon mkMaxTagOcc mk_tc_deriv_name :: TyCon -> (OccName -> OccName) -> RdrName mk_tc_deriv_name tycon occ_fun = mkAuxBinderName (tyConName tycon) occ_fun mkAuxBinderName :: Name -> (OccName -> OccName) -> RdrName -- ^ Make a top-level binder name for an auxiliary binding for a parent name -- See Note [Auxiliary binders] mkAuxBinderName parent occ_fun = mkRdrUnqual (occ_fun stable_parent_occ) where stable_parent_occ = mkOccName (occNameSpace parent_occ) stable_string stable_string | opt_PprStyle_Debug = parent_stable | otherwise = parent_stable_hash parent_stable = nameStableString parent parent_stable_hash = let Fingerprint high low = fingerprintString parent_stable in toBase62 high ++ toBase62Padded low -- See Note [Base 62 encoding 128-bit integers] parent_occ = nameOccName parent {- Note [Auxiliary binders] ~~~~~~~~~~~~~~~~~~~~~~~~ We often want to make a top-level auxiliary binding. E.g. for comparison we haev instance Ord T where compare a b = $con2tag a `compare` $con2tag b $con2tag :: T -> Int $con2tag = ...code.... Of course these top-level bindings should all have distinct name, and we are generating RdrNames here. We can't just use the TyCon or DataCon to distinguish because with standalone deriving two imported TyCons might both be called T! (See Trac #7947.) So we use package name, module name and the name of the parent (T in this example) as part of the OccName we generate for the new binding. To make the symbol names short we take a base62 hash of the full name. In the past we used the *unique* from the parent, but that's not stable across recompilations as uniques are nondeterministic. Note [DeriveFoldable with ExistentialQuantification] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Functor and Traversable instances can only be derived for data types whose last type parameter is truly universally polymorphic. For example: data T a b where T1 :: b -> T a b -- YES, b is unconstrained T2 :: Ord b => b -> T a b -- NO, b is constrained by (Ord b) T3 :: b ~ Int => b -> T a b -- NO, b is constrained by (b ~ Int) T4 :: Int -> T a Int -- NO, this is just like T3 T5 :: Ord a => a -> b -> T a b -- YES, b is unconstrained, even -- though a is existential T6 :: Int -> T Int b -- YES, b is unconstrained For Foldable instances, however, we can completely lift the constraint that the last type parameter be truly universally polymorphic. This means that T (as defined above) can have a derived Foldable instance: instance Foldable (T a) where foldr f z (T1 b) = f b z foldr f z (T2 b) = f b z foldr f z (T3 b) = f b z foldr f z (T4 b) = z foldr f z (T5 a b) = f b z foldr f z (T6 a) = z foldMap f (T1 b) = f b foldMap f (T2 b) = f b foldMap f (T3 b) = f b foldMap f (T4 b) = mempty foldMap f (T5 a b) = f b foldMap f (T6 a) = mempty In a Foldable instance, it is safe to fold over an occurrence of the last type parameter that is not truly universally polymorphic. However, there is a bit of subtlety in determining what is actually an occurrence of a type parameter. T3 and T4, as defined above, provide one example: data T a b where ... T3 :: b ~ Int => b -> T a b T4 :: Int -> T a Int ... instance Foldable (T a) where ... foldr f z (T3 b) = f b z foldr f z (T4 b) = z ... foldMap f (T3 b) = f b foldMap f (T4 b) = mempty ... Notice that the argument of T3 is folded over, whereas the argument of T4 is not. This is because we only fold over constructor arguments that syntactically mention the universally quantified type parameter of that particular data constructor. See foldDataConArgs for how this is implemented. As another example, consider the following data type. The argument of each constructor has the same type as the last type parameter: data E a where E1 :: (a ~ Int) => a -> E a E2 :: Int -> E Int E3 :: (a ~ Int) => a -> E Int E4 :: (a ~ Int) => Int -> E a Only E1's argument is an occurrence of a universally quantified type variable that is syntactically equivalent to the last type parameter, so only E1's argument will be be folded over in a derived Foldable instance. See Trac #10447 for the original discussion on this feature. Also see https://ghc.haskell.org/trac/ghc/wiki/Commentary/Compiler/DeriveFunctor for a more in-depth explanation. Note [FFoldType and functorLikeTraverse] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Deriving Functor, Foldable, and Traversable all require generating expressions which perform an operation on each argument of a data constructor depending on the argument's type. In particular, a generated operation can be different depending on whether the type mentions the last type variable of the datatype (e.g., if you have data T a = MkT a Int, then a generated foldr expresion would fold over the first argument of MkT, but not the second). This pattern is abstracted with the FFoldType datatype, which provides hooks for the user to specify how a constructor argument should be folded when it has a type with a particular "shape". The shapes are as follows (assume that a is the last type variable in a given datatype): * ft_triv: The type does not mention the last type variable at all. Examples: Int, b * ft_var: The type is syntactically equal to the last type variable. Moreover, the type appears in a covariant position (see the Deriving Functor instances section of the users' guide for an in-depth explanation of covariance vs. contravariance). Example: a (covariantly) * ft_co_var: The type is syntactically equal to the last type variable. Moreover, the type appears in a contravariant position. Example: a (contravariantly) * ft_fun: A function type which mentions the last type variable in the argument position, result position or both. Examples: a -> Int, Int -> a, Maybe a -> [a] * ft_tup: A tuple type which mentions the last type variable in at least one of its fields. The TyCon argument of ft_tup represents the particular tuple's type constructor. Examples: (a, Int), (Maybe a, [a], Either a Int) * ft_ty_app: A type is being applied to the last type parameter, where the applied type does not mention the last type parameter (if it did, it would fall under ft_bad_app). The Type argument to ft_ty_app represents the applied type. Note that functions, tuples, and foralls are distinct cases and take precedence of ft_ty_app. (For example, (Int -> a) would fall under (ft_fun Int a), not (ft_ty_app ((->) Int) a). Examples: Maybe a, Either b a * ft_bad_app: A type application uses the last type parameter in a position other than the last argument. This case is singled out because Functor, Foldable, and Traversable instances cannot be derived for datatypes containing arguments with such types. Examples: Either a Int, Const a b * ft_forall: A forall'd type mentions the last type parameter on its right- hand side (and is not quantified on the left-hand side). This case is present mostly for plumbing purposes. Example: forall b. Either b a If FFoldType describes a strategy for folding subcomponents of a Type, then functorLikeTraverse is the function that applies that strategy to the entirety of a Type, returning the final folded-up result. foldDataConArgs applies functorLikeTraverse to every argument type of a constructor, returning a list of the fold results. This makes foldDataConArgs a natural way to generate the subexpressions in a generated fmap, foldr, foldMap, or traverse definition (the subexpressions must then be combined in a method-specific fashion to form the final generated expression). Deriving Generic1 also does validity checking by looking for the last type variable in certain positions of a constructor's argument types, so it also uses foldDataConArgs. See Note [degenerate use of FFoldType] in TcGenGenerics. Note [Generated code for DeriveFoldable and DeriveTraversable] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ We adapt the algorithms for -XDeriveFoldable and -XDeriveTraversable based on that of -XDeriveFunctor. However, there an important difference between deriving the former two typeclasses and the latter one, which is best illustrated by the following scenario: data WithInt a = WithInt a Int# deriving (Functor, Foldable, Traversable) The generated code for the Functor instance is straightforward: instance Functor WithInt where fmap f (WithInt a i) = WithInt (f a) i But if we use too similar of a strategy for deriving the Foldable and Traversable instances, we end up with this code: instance Foldable WithInt where foldMap f (WithInt a i) = f a <> mempty instance Traversable WithInt where traverse f (WithInt a i) = fmap WithInt (f a) <*> pure i This is unsatisfying for two reasons: 1. The Traversable instance doesn't typecheck! Int# is of kind #, but pure expects an argument whose type is of kind *. This effectively prevents Traversable from being derived for any datatype with an unlifted argument type (Trac #11174). 2. The generated code contains superfluous expressions. By the Monoid laws, we can reduce (f a <> mempty) to (f a), and by the Applicative laws, we can reduce (fmap WithInt (f a) <*> pure i) to (fmap (\b -> WithInt b i) (f a)). We can fix both of these issues by incorporating a slight twist to the usual algorithm that we use for -XDeriveFunctor. The differences can be summarized as follows: 1. In the generated expression, we only fold over arguments whose types mention the last type parameter. Any other argument types will simply produce useless 'mempty's or 'pure's, so they can be safely ignored. 2. In the case of -XDeriveTraversable, instead of applying ConName, we apply (\b_i ... b_k -> ConName a_1 ... a_n), where * ConName has n arguments * {b_i, ..., b_k} is a subset of {a_1, ..., a_n} whose indices correspond to the arguments whose types mention the last type parameter. As a consequence, taking the difference of {a_1, ..., a_n} and {b_i, ..., b_k} yields the all the argument values of ConName whose types do not mention the last type parameter. Note that [i, ..., k] is a strictly increasing—but not necessarily consecutive—integer sequence. For example, the datatype data Foo a = Foo Int a Int a would generate the following Traversable instance: instance Traversable Foo where traverse f (Foo a1 a2 a3 a4) = fmap (\b2 b4 -> Foo a1 b2 a3 b4) (f a2) <*> f a4 Technically, this approach would also work for -XDeriveFunctor as well, but we decide not to do so because: 1. There's not much benefit to generating, e.g., ((\b -> WithInt b i) (f a)) instead of (WithInt (f a) i). 2. There would be certain datatypes for which the above strategy would generate Functor code that would fail to typecheck. For example: data Bar f a = Bar (forall f. Functor f => f a) deriving Functor With the conventional algorithm, it would generate something like: fmap f (Bar a) = Bar (fmap f a) which typechecks. But with the strategy mentioned above, it would generate: fmap f (Bar a) = (\b -> Bar b) (fmap f a) which does not typecheck, since GHC cannot unify the rank-2 type variables in the types of b and (fmap f a). -}
vTurbine/ghc
compiler/typecheck/TcGenDeriv.hs
bsd-3-clause
119,963
0
19
37,268
19,394
10,180
9,214
1,381
8
-- #hide -------------------------------------------------------------------------------- -- | -- Module : Graphics.UI.GLUT.QueryUtils -- Copyright : (c) Sven Panne 2003 -- License : BSD-style (see the file libraries/GLUT/LICENSE) -- -- Maintainer : sven_panne@yahoo.com -- Stability : provisional -- Portability : portable -- -- This is a purely internal module with utilities to query GLUT state. -- -------------------------------------------------------------------------------- module Graphics.UI.GLUT.QueryUtils ( Getter, simpleGet, layerGet, deviceGet ) where import Control.Monad ( liftM ) import Foreign.C.Types ( CInt ) import Graphics.Rendering.OpenGL.GL.BasicTypes ( GLenum ) -------------------------------------------------------------------------------- type PrimGetter = GLenum -> IO CInt type Getter a = (CInt -> a) -> GLenum -> IO a makeGetter :: PrimGetter -> Getter a makeGetter g f = liftM f . g simpleGet, layerGet, deviceGet :: Getter a simpleGet = makeGetter glutGet layerGet = makeGetter glutLayerGet deviceGet = makeGetter glutDeviceGet foreign import stdcall unsafe "HsGLUT.h glutGet" glutGet :: PrimGetter foreign import stdcall unsafe "HsGLUT.h glutLayerGet" glutLayerGet :: PrimGetter foreign import stdcall unsafe "HsGLUT.h glutDeviceGet" glutDeviceGet :: PrimGetter
OS2World/DEV-UTIL-HUGS
libraries/Graphics/UI/GLUT/QueryUtils.hs
bsd-3-clause
1,360
0
7
218
216
131
85
16
1
----------------------------------------------------------------------------- -- | -- Module : Data.Metrology.Parser -- Copyright : (C) 2014 Richard Eisenberg -- License : BSD-style (see LICENSE) -- Maintainer : Richard Eisenberg (rae@cs.brynmawr.edu) -- Stability : experimental -- Portability : non-portable -- -- This module exports functions allowing users to create their own unit -- quasiquoters to make for compact unit expressions. -- -- A typical use case is this: -- -- > $(makeQuasiQuoter "unit" [''Kilo, ''Milli] [''Meter, ''Second]) -- -- and then, /in a separate module/ (due to GHC's staging constraints) -- -- > x = 3 % [unit| m/s^2 ] -- -- The unit expressions can refer to the prefixes and units specified in -- the call to 'makeQuasiQuoter'. The spellings of the prefixes and units -- are taken from their @Show@ instances. -- -- The syntax for these expressions is like -- F#'s. There are four arithmetic operators (@*@, @/@, @^@, and juxtaposition). -- Exponentiation binds the tightest, and it allows an integer to its right -- (possibly with minus signs and parentheses). Next tightest is juxtaposition, -- which indicates multiplication. Because juxtaposition binds tighter than division, -- the expressions @m/s^2@ and @m/s s@ are equivalent. Multiplication and -- division bind the loosest and are left-associative, meaning that @m/s*s@ -- is equivalent to @(m/s)*s@, probably not what you meant. Parentheses in -- unit expressions are allowed, of course. -- -- Within a unit string (that is, a unit with an optional prefix), there may -- be ambiguity. If a unit string can be interpreted as a unit without a -- prefix, that parsing is preferred. Thus, @min@ would be minutes, not -- milli-inches (assuming appropriate prefixes and units available.) There still -- may be ambiguity between unit strings, even interpreting the string as a prefix -- and a base unit. If a unit string is amiguous in this way, it is rejected. -- For example, if we have prefixes @da@ and @d@ and units @m@ and @am@, then -- @dam@ is ambiguous like this. ----------------------------------------------------------------------------- {-# LANGUAGE TemplateHaskell, CPP #-} {-# OPTIONS_HADDOCK prune #-} module Data.Metrology.Parser ( -- * Quasiquoting interface makeQuasiQuoter, allUnits, allPrefixes, -- * Direct interface -- | The definitions below allow users to access the unit parser directly. -- The parser produces 'UnitExp's which can then be further processed as -- necessary. parseUnit, UnitExp(..), SymbolTable, mkSymbolTable, -- for internal use only parseUnitExp, parseUnitType ) where import Prelude hiding ( exp ) import Language.Haskell.TH hiding ( Pred ) import Language.Haskell.TH.Quote import Language.Haskell.TH.Desugar.Lift () -- get the Lift Name instance import Data.Maybe import Control.Monad import Text.Parse.Units import Data.Metrology import Data.Metrology.TH ---------------------------------------------------------------------- -- TH conversions ---------------------------------------------------------------------- parseUnitExp :: SymbolTable Name Name -> String -> Either String Exp parseUnitExp tab s = to_exp `liftM` parseUnit tab s -- the Either monad where to_exp Unity = ConE 'Number to_exp (Unit (Just pre) unit) = ConE '(:@) `AppE` of_type pre `AppE` of_type unit to_exp (Unit Nothing unit) = of_type unit to_exp (Mult e1 e2) = ConE '(:*) `AppE` to_exp e1 `AppE` to_exp e2 to_exp (Div e1 e2) = ConE '(:/) `AppE` to_exp e1 `AppE` to_exp e2 to_exp (Pow e i) = ConE '(:^) `AppE` to_exp e `AppE` mk_sing i of_type :: Name -> Exp of_type n = (VarE 'undefined) `SigE` (ConT n) mk_sing :: Integer -> Exp mk_sing n | n < 0 = VarE 'sPred `AppE` mk_sing (n + 1) | n > 0 = VarE 'sSucc `AppE` mk_sing (n - 1) | otherwise = VarE 'sZero parseUnitType :: SymbolTable Name Name -> String -> Either String Type parseUnitType tab s = to_type `liftM` parseUnit tab s -- the Either monad where to_type Unity = ConT ''Number to_type (Unit (Just pre) unit) = ConT ''(:@) `AppT` ConT pre `AppT` ConT unit to_type (Unit Nothing unit) = ConT unit to_type (Mult e1 e2) = ConT ''(:*) `AppT` to_type e1 `AppT` to_type e2 to_type (Div e1 e2) = ConT ''(:/) `AppT` to_type e1 `AppT` to_type e2 to_type (Pow e i) = ConT ''(:^) `AppT` to_type e `AppT` mk_z i mk_z :: Integer -> Type mk_z n | n < 0 = ConT ''Pred `AppT` mk_z (n + 1) | n > 0 = ConT ''Succ `AppT` mk_z (n - 1) | otherwise = ConT 'Zero -- single quote as it's a data constructor! ---------------------------------------------------------------------- -- QuasiQuoters ---------------------------------------------------------------------- emptyQQ :: QuasiQuoter emptyQQ = QuasiQuoter { quoteExp = \_ -> fail "No quasi-quoter for expressions" , quotePat = \_ -> fail "No quasi-quoter for patterns" , quoteType = \_ -> fail "No quasi-quoter for types" , quoteDec = \_ -> fail "No quasi-quoter for declarations" } errorQQ :: String -> QuasiQuoter errorQQ msg = QuasiQuoter { quoteExp = \_ -> fail msg , quotePat = \_ -> fail msg , quoteType = \_ -> fail msg , quoteDec = \_ -> fail msg } -- | @makeQuasiQuoter "qq" prefixes units@ makes a quasi-quoter named @qq@ -- that considers the prefixes and units provided. These are provided via -- names of the /type/ constructors, /not/ the data constructors. See the -- module documentation for more info and an example. makeQuasiQuoter :: String -> [Name] -> [Name] -> Q [Dec] makeQuasiQuoter qq_name_str prefix_names unit_names = do mapM_ checkIsType prefix_names mapM_ checkIsType unit_names qq <- [| case $sym_tab of Left err -> errorQQ err Right computed_sym_tab -> emptyQQ { quoteExp = \unit_exp -> case parseUnitExp computed_sym_tab unit_exp of Left err2 -> fail err2 Right exp -> return exp , quoteType = \unit_exp -> case parseUnitType computed_sym_tab unit_exp of Left err2 -> fail err2 Right typ -> return typ } |] return [ SigD qq_name (ConT ''QuasiQuoter) , ValD (VarP qq_name) (NormalB qq) []] where qq_name = mkName qq_name_str mk_pair :: Name -> Q Exp -- Exp is of type (String, Name) mk_pair n = [| (show (undefined :: $( return $ ConT n )), n) |] sym_tab :: Q Exp -- Exp is of type (Either String SymbolTable) sym_tab = do prefix_pairs <- mapM mk_pair prefix_names unit_pairs <- mapM mk_pair unit_names [| mkSymbolTable $( return $ ListE prefix_pairs ) $( return $ ListE unit_pairs ) |] ---------------------------------------------------------------------- -- Getting instances ---------------------------------------------------------------------- getInstanceNames :: Name -> Q [Name] getInstanceNames class_name = do ClassI _ insts <- reify class_name m_names <- forM insts $ \inst -> case inst of InstanceD #if __GLASGOW_HASKELL__ >= 711 _ #endif _ ((ConT class_name') `AppT` (ConT unit_name)) [] | class_name == class_name' -> do show_insts <- reifyInstances ''Show [ConT unit_name] case show_insts of [_show_inst] -> return $ Just unit_name _ -> return Nothing _ -> return Nothing return $ catMaybes m_names #if __GLASGOW_HASKELL__ < 709 {-# WARNING allUnits, allPrefixes "Retrieving the list of all units and prefixes in scope does not work under GHC 7.8.*. Please upgrade GHC to use these functions." #-} #endif -- | Gets a list of the names of all units with @Show@ instances in scope. -- Example usage: -- -- > $( do units <- allUnits -- > makeQuasiQuoter "unit" [] units ) -- allUnits :: Q [Name] allUnits = getInstanceNames ''Unit -- | Gets a list of the names of all unit prefixes with @Show@ instances in -- scope. Example usage: -- -- > $( do units <- allUnits -- > prefixes <- allPrefixes -- > makeQuasiQuoter "unit" prefixes units ) -- allPrefixes :: Q [Name] allPrefixes = getInstanceNames ''UnitPrefix
goldfirere/units
units/Data/Metrology/Parser.hs
bsd-3-clause
8,556
0
20
2,109
1,463
814
649
99
6
{-# LANGUAGE PackageImports #-} import Control.Applicative import "monads-tf" Control.Monad.Trans import Control.Concurrent (threadDelay) import Control.Concurrent.STM import Control.Exception import Data.Maybe import Data.Pipe import System.IO import Network import Data.Char main :: IO () main = do h <- connectTo "localhost" $ PortNumber 54492 vh <- atomically $ newTVar h fromJust <$> runPipe (input vh =$= process =$= output vh) input :: TVar Handle -> Pipe () String IO () input vh = do h <- liftIO . atomically $ readTVar vh maybe (return ()) yield =<< liftIO ((`catch` reconnect vh Nothing) . (Just <$>) $ hGetLine h) input vh output :: TVar Handle -> Pipe String () IO () output vh = do h <- liftIO . atomically $ readTVar vh await >>= \mx -> case mx of Just x -> do liftIO . (`catch` reconnect vh (Just x)) $ do hPutStrLn h x return Nothing output vh _ -> return () reconnect :: TVar Handle -> Maybe String -> IOException -> IO (Maybe String) reconnect vh x _ = do h <- connectTo "localhost" $ PortNumber 54492 maybe (return ()) (hPutStrLn h) x atomically $ writeTVar vh h return Nothing process :: Monad m => Pipe String String m () process = await >>= \mx -> case mx of Just x -> yield (map toUpper x) >> process _ -> return ()
YoshikuniJujo/xmpipe
test/s2s_model_io/s2s_cl.hs
bsd-3-clause
1,281
0
19
261
554
268
286
42
2
{-# LANGUAGE DeriveDataTypeable #-} ----------------------------------------------------------------------------- -- Some code in this file was originally from wai-1.3.0.3, made -- available under the following terms (MIT): -- -- Copyright (c) 2012 Michael Snoyman, http://www.yesodweb.com/ -- -- Permission is hereby granted, free of charge, to any person obtaining -- a copy of this software and associated documentation files (the -- "Software"), to deal in the Software without restriction, including -- without limitation the rights to use, copy, modify, merge, publish, -- distribute, sublicense, and/or sell copies of the Software, and to -- permit persons to whom the Software is furnished to do so, subject to -- the following conditions: -- -- The above copyright notice and this permission notice shall be -- included in all copies or substantial portions of the Software. -- -- THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, -- EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF -- MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND -- NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE -- LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION -- OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION -- WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. ------------------------------------------------------------------------------ module Seldom.Internal.Request ( Request(..) , RequestMeta(..) ) where import Data.Typeable (Typeable) import qualified Network.Socket as NS import qualified Data.ByteString.Char8 as B import qualified Network.HTTP.Types as H -------------------------------------------------------------------------------- data Request = Request { reqMethod :: !H.StdMethod , reqUri :: !B.ByteString , reqHttpVersion :: !H.HttpVersion , reqHeaders :: !H.RequestHeaders } deriving (Eq, Ord, Show, Typeable) data RequestMeta = RequestMeta { -- | The request message data. reqmReq :: !Request -- | Generally the host requested by the user via the Host request header. -- Backends are free to provide alternative values as necessary. This value -- should not be used to construct URLs. , reqmServerName :: !B.ByteString -- | The listening port that the server received this request on. It is -- possible for a server to listen on a non-numeric port (i.e., Unix named -- socket), in which case this value will be arbitrary. Like 'serverName', -- this value should not be used in URL construction. -- | If no query string was specified, this should be empty. This value -- /won't/ include the leading question mark. , reqQueryString :: !B.ByteString , reqmServerPort :: !B.ByteString -- | Was this request made over an SSL/TLS connection? , reqmIsSecure :: !Bool -- | The client\'s host information. , reqmRemoteHost :: !NS.SockAddr } deriving (Eq, Show, Typeable)
k0001/seldom
src/Seldom/Internal/Request.hs
bsd-3-clause
3,068
0
10
603
236
157
79
42
0
{-# LANGUAGE TupleSections #-} module Matterhorn.FilePaths ( historyFilePath , historyFileName , lastRunStateFilePath , lastRunStateFileName , configFileName , xdgName , locateConfig , xdgSyntaxDir , syntaxDirName , userEmojiJsonPath , bundledEmojiJsonPath , emojiJsonFilename , Script(..) , locateScriptPath , getAllScripts ) where import Prelude () import Matterhorn.Prelude import qualified Paths_matterhorn as Paths import Data.Text ( unpack ) import System.Directory ( doesFileExist , doesDirectoryExist , getDirectoryContents , getPermissions , executable ) import System.Environment ( getExecutablePath ) import System.Environment.XDG.BaseDir ( getUserConfigFile , getAllConfigFiles , getUserConfigDir ) import System.FilePath ( (</>), takeBaseName, takeDirectory, splitPath, joinPath ) xdgName :: String xdgName = "matterhorn" historyFileName :: FilePath historyFileName = "history.txt" lastRunStateFileName :: Text -> FilePath lastRunStateFileName teamId = "last_run_state_" ++ unpack teamId ++ ".json" configFileName :: FilePath configFileName = "config.ini" historyFilePath :: IO FilePath historyFilePath = getUserConfigFile xdgName historyFileName lastRunStateFilePath :: Text -> IO FilePath lastRunStateFilePath teamId = getUserConfigFile xdgName (lastRunStateFileName teamId) -- | Get the XDG path to the user-specific syntax definition directory. -- The path does not necessarily exist. xdgSyntaxDir :: IO FilePath xdgSyntaxDir = (</> syntaxDirName) <$> getUserConfigDir xdgName -- | Get the XDG path to the user-specific emoji JSON file. The path -- does not necessarily exist. userEmojiJsonPath :: IO FilePath userEmojiJsonPath = (</> emojiJsonFilename) <$> getUserConfigDir xdgName -- | Get the emoji JSON path relative to the development binary location -- or the release binary location. bundledEmojiJsonPath :: IO FilePath bundledEmojiJsonPath = do selfPath <- getExecutablePath let distDir = "dist-newstyle/" pathBits = splitPath selfPath adjacentEmojiJsonPath <- do let path = takeDirectory selfPath </> emojiDirName </> emojiJsonFilename exists <- doesFileExist path return $ if exists then Just path else Nothing cabalEmojiJsonPath <- Paths.getDataFileName $ emojiDirName </> emojiJsonFilename return $ if distDir `elem` pathBits then -- We're in development, so use the development -- executable path to locate the emoji path in the -- development tree. (joinPath $ takeWhile (/= distDir) pathBits) </> emojiDirName </> emojiJsonFilename else -- In this case we assume the binary is being run from -- a release, in which case the syntax directory is a -- sibling of the executable path. If it does not exist -- we fall back to the cabal data files path discovered -- via Paths.getDataFileName. fromMaybe cabalEmojiJsonPath adjacentEmojiJsonPath emojiJsonFilename :: FilePath emojiJsonFilename = "emoji.json" emojiDirName :: FilePath emojiDirName = "emoji" syntaxDirName :: FilePath syntaxDirName = "syntax" -- | Find a specified configuration file by looking in all of the -- supported locations. locateConfig :: FilePath -> IO (Maybe FilePath) locateConfig filename = do xdgLocations <- getAllConfigFiles xdgName filename let confLocations = ["./" <> filename] ++ xdgLocations ++ ["/etc/matterhorn/" <> filename] results <- forM confLocations $ \fp -> (fp,) <$> doesFileExist fp return $ listToMaybe $ fst <$> filter snd results scriptDirName :: FilePath scriptDirName = "scripts" data Script = ScriptPath FilePath | NonexecScriptPath FilePath | ScriptNotFound deriving (Eq, Read, Show) toScript :: FilePath -> IO (Script) toScript fp = do perm <- getPermissions fp return $ if executable perm then ScriptPath fp else NonexecScriptPath fp isExecutable :: FilePath -> IO Bool isExecutable fp = do perm <- getPermissions fp return (executable perm) locateScriptPath :: FilePath -> IO Script locateScriptPath name | head name == '.' = return ScriptNotFound | otherwise = do xdgLocations <- getAllConfigFiles xdgName scriptDirName let cmdLocations = [ xdgLoc ++ "/" ++ name | xdgLoc <- xdgLocations ] ++ [ "/etc/matterhorn/scripts/" <> name ] existingFiles <- filterM doesFileExist cmdLocations executables <- mapM toScript existingFiles return $ case executables of (path:_) -> path _ -> ScriptNotFound -- | This returns a list of valid scripts, and a list of non-executable -- scripts. getAllScripts :: IO ([FilePath], [FilePath]) getAllScripts = do xdgLocations <- getAllConfigFiles xdgName scriptDirName let cmdLocations = xdgLocations ++ ["/etc/matterhorn/scripts"] let getCommands dir = do exists <- doesDirectoryExist dir if exists then map ((dir ++ "/") ++) `fmap` getDirectoryContents dir else return [] let isNotHidden f = case f of ('.':_) -> False [] -> False _ -> True allScripts <- concat `fmap` mapM getCommands cmdLocations execs <- filterM isExecutable allScripts nonexecs <- filterM (fmap not . isExecutable) allScripts return ( filter isNotHidden $ map takeBaseName execs , filter isNotHidden $ map takeBaseName nonexecs )
matterhorn-chat/matterhorn
src/Matterhorn/FilePaths.hs
bsd-3-clause
5,779
0
17
1,496
1,176
618
558
123
4
{- (c) The University of Glasgow 2006 (c) The GRASP/AQUA Project, Glasgow University, 1992-1998 -} {-# LANGUAGE RankNTypes #-} {-# LANGUAGE TypeFamilies #-} {-# LANGUAGE FlexibleContexts #-} {-# LANGUAGE TypeSynonymInstances #-} {-# LANGUAGE FlexibleInstances #-} {-# LANGUAGE UndecidableInstances #-} module TrieMap( CoreMap, emptyCoreMap, extendCoreMap, lookupCoreMap, foldCoreMap, TypeMap, emptyTypeMap, extendTypeMap, lookupTypeMap, foldTypeMap, LooseTypeMap, MaybeMap, ListMap, TrieMap(..), insertTM, deleteTM ) where import CoreSyn import Coercion import Literal import Name import Type import TyCoRep import Var import UniqDFM import Unique( Unique ) import FastString(FastString) import qualified Data.Map as Map import qualified Data.IntMap as IntMap import VarEnv import NameEnv import Outputable import Control.Monad( (>=>) ) {- This module implements TrieMaps, which are finite mappings whose key is a structured value like a CoreExpr or Type. The code is very regular and boilerplate-like, but there is some neat handling of *binders*. In effect they are deBruijn numbered on the fly. The regular pattern for handling TrieMaps on data structures was first described (to my knowledge) in Connelly and Morris's 1995 paper "A generalization of the Trie Data Structure"; there is also an accessible description of the idea in Okasaki's book "Purely Functional Data Structures", Section 10.3.2 ************************************************************************ * * The TrieMap class * * ************************************************************************ -} type XT a = Maybe a -> Maybe a -- How to alter a non-existent elt (Nothing) -- or an existing elt (Just) class TrieMap m where type Key m :: * emptyTM :: m a lookupTM :: forall b. Key m -> m b -> Maybe b alterTM :: forall b. Key m -> XT b -> m b -> m b mapTM :: (a->b) -> m a -> m b foldTM :: (a -> b -> b) -> m a -> b -> b -- The unusual argument order here makes -- it easy to compose calls to foldTM; -- see for example fdE below insertTM :: TrieMap m => Key m -> a -> m a -> m a insertTM k v m = alterTM k (\_ -> Just v) m deleteTM :: TrieMap m => Key m -> m a -> m a deleteTM k m = alterTM k (\_ -> Nothing) m ---------------------- -- Recall that -- Control.Monad.(>=>) :: (a -> Maybe b) -> (b -> Maybe c) -> a -> Maybe c (>.>) :: (a -> b) -> (b -> c) -> a -> c -- Reverse function composition (do f first, then g) infixr 1 >.> (f >.> g) x = g (f x) infixr 1 |>, |>> (|>) :: a -> (a->b) -> b -- Reverse application x |> f = f x ---------------------- (|>>) :: TrieMap m2 => (XT (m2 a) -> m1 (m2 a) -> m1 (m2 a)) -> (m2 a -> m2 a) -> m1 (m2 a) -> m1 (m2 a) (|>>) f g = f (Just . g . deMaybe) deMaybe :: TrieMap m => Maybe (m a) -> m a deMaybe Nothing = emptyTM deMaybe (Just m) = m {- ************************************************************************ * * IntMaps * * ************************************************************************ -} instance TrieMap IntMap.IntMap where type Key IntMap.IntMap = Int emptyTM = IntMap.empty lookupTM k m = IntMap.lookup k m alterTM = xtInt foldTM k m z = IntMap.fold k z m mapTM f m = IntMap.map f m xtInt :: Int -> XT a -> IntMap.IntMap a -> IntMap.IntMap a xtInt k f m = IntMap.alter f k m instance Ord k => TrieMap (Map.Map k) where type Key (Map.Map k) = k emptyTM = Map.empty lookupTM = Map.lookup alterTM k f m = Map.alter f k m foldTM k m z = Map.fold k z m mapTM f m = Map.map f m {- Note [foldTM determinism] ~~~~~~~~~~~~~~~~~~~~~~~~~ We want foldTM to be deterministic, which is why we have an instance of TrieMap for UniqDFM, but not for UniqFM. Here's an example of some things that go wrong if foldTM is nondeterministic. Consider: f a b = return (a <> b) Depending on the order that the typechecker generates constraints you get either: f :: (Monad m, Monoid a) => a -> a -> m a or: f :: (Monoid a, Monad m) => a -> a -> m a The generated code will be different after desugaring as the dictionaries will be bound in different orders, leading to potential ABI incompatibility. One way to solve this would be to notice that the typeclasses could be sorted alphabetically. Unfortunately that doesn't quite work with this example: f a b = let x = a <> a; y = b <> b in x where you infer: f :: (Monoid m, Monoid m1) => m1 -> m -> m1 or: f :: (Monoid m1, Monoid m) => m1 -> m -> m1 Here you could decide to take the order of the type variables in the type according to depth first traversal and use it to order the constraints. The real trouble starts when the user enables incoherent instances and the compiler has to make an arbitrary choice. Consider: class T a b where go :: a -> b -> String instance (Show b) => T Int b where go a b = show a ++ show b instance (Show a) => T a Bool where go a b = show a ++ show b f = go 10 True GHC is free to choose either dictionary to implement f, but for the sake of determinism we'd like it to be consistent when compiling the same sources with the same flags. inert_dicts :: DictMap is implemented with a TrieMap. In getUnsolvedInerts it gets converted to a bag of (Wanted) Cts using a fold. Then in solve_simple_wanteds it's merged with other WantedConstraints. We want the conversion to a bag to be deterministic. For that purpose we use UniqDFM instead of UniqFM to implement the TrieMap. See Note [Deterministic UniqFM] in UniqDFM for more details on how it's made deterministic. -} instance TrieMap UniqDFM where type Key UniqDFM = Unique emptyTM = emptyUDFM lookupTM k m = lookupUDFM m k alterTM k f m = alterUDFM f m k foldTM k m z = foldUDFM k z m mapTM f m = mapUDFM f m {- ************************************************************************ * * Maybes * * ************************************************************************ If m is a map from k -> val then (MaybeMap m) is a map from (Maybe k) -> val -} data MaybeMap m a = MM { mm_nothing :: Maybe a, mm_just :: m a } instance TrieMap m => TrieMap (MaybeMap m) where type Key (MaybeMap m) = Maybe (Key m) emptyTM = MM { mm_nothing = Nothing, mm_just = emptyTM } lookupTM = lkMaybe lookupTM alterTM = xtMaybe alterTM foldTM = fdMaybe mapTM = mapMb mapMb :: TrieMap m => (a->b) -> MaybeMap m a -> MaybeMap m b mapMb f (MM { mm_nothing = mn, mm_just = mj }) = MM { mm_nothing = fmap f mn, mm_just = mapTM f mj } lkMaybe :: (forall b. k -> m b -> Maybe b) -> Maybe k -> MaybeMap m a -> Maybe a lkMaybe _ Nothing = mm_nothing lkMaybe lk (Just x) = mm_just >.> lk x xtMaybe :: (forall b. k -> XT b -> m b -> m b) -> Maybe k -> XT a -> MaybeMap m a -> MaybeMap m a xtMaybe _ Nothing f m = m { mm_nothing = f (mm_nothing m) } xtMaybe tr (Just x) f m = m { mm_just = mm_just m |> tr x f } fdMaybe :: TrieMap m => (a -> b -> b) -> MaybeMap m a -> b -> b fdMaybe k m = foldMaybe k (mm_nothing m) . foldTM k (mm_just m) {- ************************************************************************ * * Lists * * ************************************************************************ -} data ListMap m a = LM { lm_nil :: Maybe a , lm_cons :: m (ListMap m a) } instance TrieMap m => TrieMap (ListMap m) where type Key (ListMap m) = [Key m] emptyTM = LM { lm_nil = Nothing, lm_cons = emptyTM } lookupTM = lkList lookupTM alterTM = xtList alterTM foldTM = fdList mapTM = mapList mapList :: TrieMap m => (a->b) -> ListMap m a -> ListMap m b mapList f (LM { lm_nil = mnil, lm_cons = mcons }) = LM { lm_nil = fmap f mnil, lm_cons = mapTM (mapTM f) mcons } lkList :: TrieMap m => (forall b. k -> m b -> Maybe b) -> [k] -> ListMap m a -> Maybe a lkList _ [] = lm_nil lkList lk (x:xs) = lm_cons >.> lk x >=> lkList lk xs xtList :: TrieMap m => (forall b. k -> XT b -> m b -> m b) -> [k] -> XT a -> ListMap m a -> ListMap m a xtList _ [] f m = m { lm_nil = f (lm_nil m) } xtList tr (x:xs) f m = m { lm_cons = lm_cons m |> tr x |>> xtList tr xs f } fdList :: forall m a b. TrieMap m => (a -> b -> b) -> ListMap m a -> b -> b fdList k m = foldMaybe k (lm_nil m) . foldTM (fdList k) (lm_cons m) foldMaybe :: (a -> b -> b) -> Maybe a -> b -> b foldMaybe _ Nothing b = b foldMaybe k (Just a) b = k a b {- ************************************************************************ * * Basic maps * * ************************************************************************ -} lkDNamed :: NamedThing n => n -> DNameEnv a -> Maybe a lkDNamed n env = lookupDNameEnv env (getName n) xtDNamed :: NamedThing n => n -> XT a -> DNameEnv a -> DNameEnv a xtDNamed tc f m = alterDNameEnv f m (getName tc) ------------------------ type LiteralMap a = Map.Map Literal a emptyLiteralMap :: LiteralMap a emptyLiteralMap = emptyTM lkLit :: Literal -> LiteralMap a -> Maybe a lkLit = lookupTM xtLit :: Literal -> XT a -> LiteralMap a -> LiteralMap a xtLit = alterTM {- ************************************************************************ * * GenMap * * ************************************************************************ Note [Compressed TrieMap] ~~~~~~~~~~~~~~~~~~~~~~~~~ The GenMap constructor augments TrieMaps with leaf compression. This helps solve the performance problem detailed in #9960: suppose we have a handful H of entries in a TrieMap, each with a very large key, size K. If you fold over such a TrieMap you'd expect time O(H). That would certainly be true of an association list! But with TrieMap we actually have to navigate down a long singleton structure to get to the elements, so it takes time O(K*H). This can really hurt on many type-level computation benchmarks: see for example T9872d. The point of a TrieMap is that you need to navigate to the point where only one key remains, and then things should be fast. So the point of a SingletonMap is that, once we are down to a single (key,value) pair, we stop and just use SingletonMap. 'EmptyMap' provides an even more basic (but essential) optimization: if there is nothing in the map, don't bother building out the (possibly infinite) recursive TrieMap structure! -} data GenMap m a = EmptyMap | SingletonMap (Key m) a | MultiMap (m a) instance (Outputable a, Outputable (m a)) => Outputable (GenMap m a) where ppr EmptyMap = text "Empty map" ppr (SingletonMap _ v) = text "Singleton map" <+> ppr v ppr (MultiMap m) = ppr m -- TODO undecidable instance instance (Eq (Key m), TrieMap m) => TrieMap (GenMap m) where type Key (GenMap m) = Key m emptyTM = EmptyMap lookupTM = lkG alterTM = xtG foldTM = fdG mapTM = mapG -- NB: Be careful about RULES and type families (#5821). So we should make sure -- to specify @Key TypeMapX@ (and not @DeBruijn Type@, the reduced form) {-# SPECIALIZE lkG :: Key TypeMapX -> TypeMapG a -> Maybe a #-} {-# SPECIALIZE lkG :: Key CoercionMapX -> CoercionMapG a -> Maybe a #-} {-# SPECIALIZE lkG :: Key CoreMapX -> CoreMapG a -> Maybe a #-} lkG :: (Eq (Key m), TrieMap m) => Key m -> GenMap m a -> Maybe a lkG _ EmptyMap = Nothing lkG k (SingletonMap k' v') | k == k' = Just v' | otherwise = Nothing lkG k (MultiMap m) = lookupTM k m {-# SPECIALIZE xtG :: Key TypeMapX -> XT a -> TypeMapG a -> TypeMapG a #-} {-# SPECIALIZE xtG :: Key CoercionMapX -> XT a -> CoercionMapG a -> CoercionMapG a #-} {-# SPECIALIZE xtG :: Key CoreMapX -> XT a -> CoreMapG a -> CoreMapG a #-} xtG :: (Eq (Key m), TrieMap m) => Key m -> XT a -> GenMap m a -> GenMap m a xtG k f EmptyMap = case f Nothing of Just v -> SingletonMap k v Nothing -> EmptyMap xtG k f m@(SingletonMap k' v') | k' == k -- The new key matches the (single) key already in the tree. Hence, -- apply @f@ to @Just v'@ and build a singleton or empty map depending -- on the 'Just'/'Nothing' response respectively. = case f (Just v') of Just v'' -> SingletonMap k' v'' Nothing -> EmptyMap | otherwise -- We've hit a singleton tree for a different key than the one we are -- searching for. Hence apply @f@ to @Nothing@. If result is @Nothing@ then -- we can just return the old map. If not, we need a map with *two* -- entries. The easiest way to do that is to insert two items into an empty -- map of type @m a@. = case f Nothing of Nothing -> m Just v -> emptyTM |> alterTM k' (const (Just v')) >.> alterTM k (const (Just v)) >.> MultiMap xtG k f (MultiMap m) = MultiMap (alterTM k f m) {-# SPECIALIZE mapG :: (a -> b) -> TypeMapG a -> TypeMapG b #-} {-# SPECIALIZE mapG :: (a -> b) -> CoercionMapG a -> CoercionMapG b #-} {-# SPECIALIZE mapG :: (a -> b) -> CoreMapG a -> CoreMapG b #-} mapG :: TrieMap m => (a -> b) -> GenMap m a -> GenMap m b mapG _ EmptyMap = EmptyMap mapG f (SingletonMap k v) = SingletonMap k (f v) mapG f (MultiMap m) = MultiMap (mapTM f m) {-# SPECIALIZE fdG :: (a -> b -> b) -> TypeMapG a -> b -> b #-} {-# SPECIALIZE fdG :: (a -> b -> b) -> CoercionMapG a -> b -> b #-} {-# SPECIALIZE fdG :: (a -> b -> b) -> CoreMapG a -> b -> b #-} fdG :: TrieMap m => (a -> b -> b) -> GenMap m a -> b -> b fdG _ EmptyMap = \z -> z fdG k (SingletonMap _ v) = \z -> k v z fdG k (MultiMap m) = foldTM k m {- ************************************************************************ * * CoreMap * * ************************************************************************ Note [Binders] ~~~~~~~~~~~~~~ * In general we check binders as late as possible because types are less likely to differ than expression structure. That's why cm_lam :: CoreMapG (TypeMapG a) rather than cm_lam :: TypeMapG (CoreMapG a) * We don't need to look at the type of some binders, notalby - the case binder in (Case _ b _ _) - the binders in an alternative because they are totally fixed by the context Note [Empty case alternatives] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ * For a key (Case e b ty (alt:alts)) we don't need to look the return type 'ty', because every alternative has that type. * For a key (Case e b ty []) we MUST look at the return type 'ty', because otherwise (Case (error () "urk") _ Int []) would compare equal to (Case (error () "urk") _ Bool []) which is utterly wrong (Trac #6097) We could compare the return type regardless, but the wildly common case is that it's unnecessary, so we have two fields (cm_case and cm_ecase) for the two possibilities. Only cm_ecase looks at the type. See also Note [Empty case alternatives] in CoreSyn. -} -- | @CoreMap a@ is a map from 'CoreExpr' to @a@. If you are a client, this -- is the type you want. newtype CoreMap a = CoreMap (CoreMapG a) instance TrieMap CoreMap where type Key CoreMap = CoreExpr emptyTM = CoreMap emptyTM lookupTM k (CoreMap m) = lookupTM (deBruijnize k) m alterTM k f (CoreMap m) = CoreMap (alterTM (deBruijnize k) f m) foldTM k (CoreMap m) = foldTM k m mapTM f (CoreMap m) = CoreMap (mapTM f m) -- | @CoreMapG a@ is a map from @DeBruijn CoreExpr@ to @a@. The extended -- key makes it suitable for recursive traversal, since it can track binders, -- but it is strictly internal to this module. If you are including a 'CoreMap' -- inside another 'TrieMap', this is the type you want. type CoreMapG = GenMap CoreMapX -- | @CoreMapX a@ is the base map from @DeBruijn CoreExpr@ to @a@, but without -- the 'GenMap' optimization. data CoreMapX a = CM { cm_var :: VarMap a , cm_lit :: LiteralMap a , cm_co :: CoercionMapG a , cm_type :: TypeMapG a , cm_cast :: CoreMapG (CoercionMapG a) , cm_tick :: CoreMapG (TickishMap a) , cm_app :: CoreMapG (CoreMapG a) , cm_lam :: CoreMapG (BndrMap a) -- Note [Binders] , cm_letn :: CoreMapG (CoreMapG (BndrMap a)) , cm_letr :: ListMap CoreMapG (CoreMapG (ListMap BndrMap a)) , cm_case :: CoreMapG (ListMap AltMap a) , cm_ecase :: CoreMapG (TypeMapG a) -- Note [Empty case alternatives] } instance Eq (DeBruijn CoreExpr) where D env1 e1 == D env2 e2 = go e1 e2 where go (Var v1) (Var v2) = case (lookupCME env1 v1, lookupCME env2 v2) of (Just b1, Just b2) -> b1 == b2 (Nothing, Nothing) -> v1 == v2 _ -> False go (Lit lit1) (Lit lit2) = lit1 == lit2 go (Type t1) (Type t2) = D env1 t1 == D env2 t2 go (Coercion co1) (Coercion co2) = D env1 co1 == D env2 co2 go (Cast e1 co1) (Cast e2 co2) = D env1 co1 == D env2 co2 && go e1 e2 go (App f1 a1) (App f2 a2) = go f1 f2 && go a1 a2 -- This seems a bit dodgy, see 'eqTickish' go (Tick n1 e1) (Tick n2 e2) = n1 == n2 && go e1 e2 go (Lam b1 e1) (Lam b2 e2) = D env1 (varType b1) == D env2 (varType b2) && D (extendCME env1 b1) e1 == D (extendCME env2 b2) e2 go (Let (NonRec v1 r1) e1) (Let (NonRec v2 r2) e2) = go r1 r2 && D (extendCME env1 v1) e1 == D (extendCME env2 v2) e2 go (Let (Rec ps1) e1) (Let (Rec ps2) e2) = length ps1 == length ps2 && D env1' rs1 == D env2' rs2 && D env1' e1 == D env2' e2 where (bs1,rs1) = unzip ps1 (bs2,rs2) = unzip ps2 env1' = extendCMEs env1 bs1 env2' = extendCMEs env2 bs2 go (Case e1 b1 t1 a1) (Case e2 b2 t2 a2) | null a1 -- See Note [Empty case alternatives] = null a2 && go e1 e2 && D env1 t1 == D env2 t2 | otherwise = go e1 e2 && D (extendCME env1 b1) a1 == D (extendCME env2 b2) a2 go _ _ = False emptyE :: CoreMapX a emptyE = CM { cm_var = emptyTM, cm_lit = emptyLiteralMap , cm_co = emptyTM, cm_type = emptyTM , cm_cast = emptyTM, cm_app = emptyTM , cm_lam = emptyTM, cm_letn = emptyTM , cm_letr = emptyTM, cm_case = emptyTM , cm_ecase = emptyTM, cm_tick = emptyTM } instance TrieMap CoreMapX where type Key CoreMapX = DeBruijn CoreExpr emptyTM = emptyE lookupTM = lkE alterTM = xtE foldTM = fdE mapTM = mapE -------------------------- mapE :: (a->b) -> CoreMapX a -> CoreMapX b mapE f (CM { cm_var = cvar, cm_lit = clit , cm_co = cco, cm_type = ctype , cm_cast = ccast , cm_app = capp , cm_lam = clam, cm_letn = cletn , cm_letr = cletr, cm_case = ccase , cm_ecase = cecase, cm_tick = ctick }) = CM { cm_var = mapTM f cvar, cm_lit = mapTM f clit , cm_co = mapTM f cco, cm_type = mapTM f ctype , cm_cast = mapTM (mapTM f) ccast, cm_app = mapTM (mapTM f) capp , cm_lam = mapTM (mapTM f) clam, cm_letn = mapTM (mapTM (mapTM f)) cletn , cm_letr = mapTM (mapTM (mapTM f)) cletr, cm_case = mapTM (mapTM f) ccase , cm_ecase = mapTM (mapTM f) cecase, cm_tick = mapTM (mapTM f) ctick } -------------------------- lookupCoreMap :: CoreMap a -> CoreExpr -> Maybe a lookupCoreMap cm e = lookupTM e cm extendCoreMap :: CoreMap a -> CoreExpr -> a -> CoreMap a extendCoreMap m e v = alterTM e (\_ -> Just v) m foldCoreMap :: (a -> b -> b) -> b -> CoreMap a -> b foldCoreMap k z m = foldTM k m z emptyCoreMap :: CoreMap a emptyCoreMap = emptyTM instance Outputable a => Outputable (CoreMap a) where ppr m = text "CoreMap elts" <+> ppr (foldTM (:) m []) ------------------------- fdE :: (a -> b -> b) -> CoreMapX a -> b -> b fdE k m = foldTM k (cm_var m) . foldTM k (cm_lit m) . foldTM k (cm_co m) . foldTM k (cm_type m) . foldTM (foldTM k) (cm_cast m) . foldTM (foldTM k) (cm_tick m) . foldTM (foldTM k) (cm_app m) . foldTM (foldTM k) (cm_lam m) . foldTM (foldTM (foldTM k)) (cm_letn m) . foldTM (foldTM (foldTM k)) (cm_letr m) . foldTM (foldTM k) (cm_case m) . foldTM (foldTM k) (cm_ecase m) -- lkE: lookup in trie for expressions lkE :: DeBruijn CoreExpr -> CoreMapX a -> Maybe a lkE (D env expr) cm = go expr cm where go (Var v) = cm_var >.> lkVar env v go (Lit l) = cm_lit >.> lkLit l go (Type t) = cm_type >.> lkG (D env t) go (Coercion c) = cm_co >.> lkG (D env c) go (Cast e c) = cm_cast >.> lkG (D env e) >=> lkG (D env c) go (Tick tickish e) = cm_tick >.> lkG (D env e) >=> lkTickish tickish go (App e1 e2) = cm_app >.> lkG (D env e2) >=> lkG (D env e1) go (Lam v e) = cm_lam >.> lkG (D (extendCME env v) e) >=> lkBndr env v go (Let (NonRec b r) e) = cm_letn >.> lkG (D env r) >=> lkG (D (extendCME env b) e) >=> lkBndr env b go (Let (Rec prs) e) = let (bndrs,rhss) = unzip prs env1 = extendCMEs env bndrs in cm_letr >.> lkList (lkG . D env1) rhss >=> lkG (D env1 e) >=> lkList (lkBndr env1) bndrs go (Case e b ty as) -- See Note [Empty case alternatives] | null as = cm_ecase >.> lkG (D env e) >=> lkG (D env ty) | otherwise = cm_case >.> lkG (D env e) >=> lkList (lkA (extendCME env b)) as xtE :: DeBruijn CoreExpr -> XT a -> CoreMapX a -> CoreMapX a xtE (D env (Var v)) f m = m { cm_var = cm_var m |> xtVar env v f } xtE (D env (Type t)) f m = m { cm_type = cm_type m |> xtG (D env t) f } xtE (D env (Coercion c)) f m = m { cm_co = cm_co m |> xtG (D env c) f } xtE (D _ (Lit l)) f m = m { cm_lit = cm_lit m |> xtLit l f } xtE (D env (Cast e c)) f m = m { cm_cast = cm_cast m |> xtG (D env e) |>> xtG (D env c) f } xtE (D env (Tick t e)) f m = m { cm_tick = cm_tick m |> xtG (D env e) |>> xtTickish t f } xtE (D env (App e1 e2)) f m = m { cm_app = cm_app m |> xtG (D env e2) |>> xtG (D env e1) f } xtE (D env (Lam v e)) f m = m { cm_lam = cm_lam m |> xtG (D (extendCME env v) e) |>> xtBndr env v f } xtE (D env (Let (NonRec b r) e)) f m = m { cm_letn = cm_letn m |> xtG (D (extendCME env b) e) |>> xtG (D env r) |>> xtBndr env b f } xtE (D env (Let (Rec prs) e)) f m = m { cm_letr = let (bndrs,rhss) = unzip prs env1 = extendCMEs env bndrs in cm_letr m |> xtList (xtG . D env1) rhss |>> xtG (D env1 e) |>> xtList (xtBndr env1) bndrs f } xtE (D env (Case e b ty as)) f m | null as = m { cm_ecase = cm_ecase m |> xtG (D env e) |>> xtG (D env ty) f } | otherwise = m { cm_case = cm_case m |> xtG (D env e) |>> let env1 = extendCME env b in xtList (xtA env1) as f } -- TODO: this seems a bit dodgy, see 'eqTickish' type TickishMap a = Map.Map (Tickish Id) a lkTickish :: Tickish Id -> TickishMap a -> Maybe a lkTickish = lookupTM xtTickish :: Tickish Id -> XT a -> TickishMap a -> TickishMap a xtTickish = alterTM ------------------------ data AltMap a -- A single alternative = AM { am_deflt :: CoreMapG a , am_data :: DNameEnv (CoreMapG a) , am_lit :: LiteralMap (CoreMapG a) } instance TrieMap AltMap where type Key AltMap = CoreAlt emptyTM = AM { am_deflt = emptyTM , am_data = emptyDNameEnv , am_lit = emptyLiteralMap } lookupTM = lkA emptyCME alterTM = xtA emptyCME foldTM = fdA mapTM = mapA instance Eq (DeBruijn CoreAlt) where D env1 a1 == D env2 a2 = go a1 a2 where go (DEFAULT, _, rhs1) (DEFAULT, _, rhs2) = D env1 rhs1 == D env2 rhs2 go (LitAlt lit1, _, rhs1) (LitAlt lit2, _, rhs2) = lit1 == lit2 && D env1 rhs1 == D env2 rhs2 go (DataAlt dc1, bs1, rhs1) (DataAlt dc2, bs2, rhs2) = dc1 == dc2 && D (extendCMEs env1 bs1) rhs1 == D (extendCMEs env2 bs2) rhs2 go _ _ = False mapA :: (a->b) -> AltMap a -> AltMap b mapA f (AM { am_deflt = adeflt, am_data = adata, am_lit = alit }) = AM { am_deflt = mapTM f adeflt , am_data = mapTM (mapTM f) adata , am_lit = mapTM (mapTM f) alit } lkA :: CmEnv -> CoreAlt -> AltMap a -> Maybe a lkA env (DEFAULT, _, rhs) = am_deflt >.> lkG (D env rhs) lkA env (LitAlt lit, _, rhs) = am_lit >.> lkLit lit >=> lkG (D env rhs) lkA env (DataAlt dc, bs, rhs) = am_data >.> lkDNamed dc >=> lkG (D (extendCMEs env bs) rhs) xtA :: CmEnv -> CoreAlt -> XT a -> AltMap a -> AltMap a xtA env (DEFAULT, _, rhs) f m = m { am_deflt = am_deflt m |> xtG (D env rhs) f } xtA env (LitAlt l, _, rhs) f m = m { am_lit = am_lit m |> xtLit l |>> xtG (D env rhs) f } xtA env (DataAlt d, bs, rhs) f m = m { am_data = am_data m |> xtDNamed d |>> xtG (D (extendCMEs env bs) rhs) f } fdA :: (a -> b -> b) -> AltMap a -> b -> b fdA k m = foldTM k (am_deflt m) . foldTM (foldTM k) (am_data m) . foldTM (foldTM k) (am_lit m) {- ************************************************************************ * * Coercions * * ************************************************************************ -} -- We should really never care about the contents of a coercion. Instead, -- just look up the coercion's type. newtype CoercionMap a = CoercionMap (CoercionMapG a) instance TrieMap CoercionMap where type Key CoercionMap = Coercion emptyTM = CoercionMap emptyTM lookupTM k (CoercionMap m) = lookupTM (deBruijnize k) m alterTM k f (CoercionMap m) = CoercionMap (alterTM (deBruijnize k) f m) foldTM k (CoercionMap m) = foldTM k m mapTM f (CoercionMap m) = CoercionMap (mapTM f m) type CoercionMapG = GenMap CoercionMapX newtype CoercionMapX a = CoercionMapX (TypeMapX a) instance TrieMap CoercionMapX where type Key CoercionMapX = DeBruijn Coercion emptyTM = CoercionMapX emptyTM lookupTM = lkC alterTM = xtC foldTM f (CoercionMapX core_tm) = foldTM f core_tm mapTM f (CoercionMapX core_tm) = CoercionMapX (mapTM f core_tm) instance Eq (DeBruijn Coercion) where D env1 co1 == D env2 co2 = D env1 (coercionType co1) == D env2 (coercionType co2) lkC :: DeBruijn Coercion -> CoercionMapX a -> Maybe a lkC (D env co) (CoercionMapX core_tm) = lkT (D env $ coercionType co) core_tm xtC :: DeBruijn Coercion -> XT a -> CoercionMapX a -> CoercionMapX a xtC (D env co) f (CoercionMapX m) = CoercionMapX (xtT (D env $ coercionType co) f m) {- ************************************************************************ * * Types * * ************************************************************************ -} -- | @TypeMapG a@ is a map from @DeBruijn Type@ to @a@. The extended -- key makes it suitable for recursive traversal, since it can track binders, -- but it is strictly internal to this module. If you are including a 'TypeMap' -- inside another 'TrieMap', this is the type you want. Note that this -- lookup does not do a kind-check. Thus, all keys in this map must have -- the same kind. type TypeMapG = GenMap TypeMapX -- | @TypeMapX a@ is the base map from @DeBruijn Type@ to @a@, but without the -- 'GenMap' optimization. data TypeMapX a = TM { tm_var :: VarMap a , tm_app :: TypeMapG (TypeMapG a) , tm_tycon :: DNameEnv a , tm_forall :: TypeMapG (BndrMap a) -- See Note [Binders] , tm_tylit :: TyLitMap a , tm_coerce :: Maybe a } -- Note that there is no tyconapp case; see Note [Equality on AppTys] in Type -- | squeeze out any synonyms, convert Constraint to *, and change TyConApps -- to nested AppTys. Why the last one? See Note [Equality on AppTys] in Type trieMapView :: Type -> Maybe Type trieMapView ty | Just ty' <- coreViewOneStarKind ty = Just ty' trieMapView (TyConApp tc tys@(_:_)) = Just $ foldl AppTy (TyConApp tc []) tys trieMapView (ForAllTy (Anon arg) res) = Just ((TyConApp funTyCon [] `AppTy` arg) `AppTy` res) trieMapView _ = Nothing instance TrieMap TypeMapX where type Key TypeMapX = DeBruijn Type emptyTM = emptyT lookupTM = lkT alterTM = xtT foldTM = fdT mapTM = mapT instance Eq (DeBruijn Type) where env_t@(D env t) == env_t'@(D env' t') | Just new_t <- coreViewOneStarKind t = D env new_t == env_t' | Just new_t' <- coreViewOneStarKind t' = env_t == D env' new_t' | otherwise = case (t, t') of (CastTy t1 _, _) -> D env t1 == D env t' (_, CastTy t1' _) -> D env t == D env t1' (TyVarTy v, TyVarTy v') -> case (lookupCME env v, lookupCME env' v') of (Just bv, Just bv') -> bv == bv' (Nothing, Nothing) -> v == v' _ -> False -- See Note [Equality on AppTys] in Type (AppTy t1 t2, s) | Just (t1', t2') <- repSplitAppTy_maybe s -> D env t1 == D env' t1' && D env t2 == D env' t2' (s, AppTy t1' t2') | Just (t1, t2) <- repSplitAppTy_maybe s -> D env t1 == D env' t1' && D env t2 == D env' t2' (ForAllTy (Anon t1) t2, ForAllTy (Anon t1') t2') -> D env t1 == D env' t1' && D env t2 == D env' t2' (TyConApp tc tys, TyConApp tc' tys') -> tc == tc' && D env tys == D env' tys' (LitTy l, LitTy l') -> l == l' (ForAllTy (Named tv _) ty, ForAllTy (Named tv' _) ty') -> D env (tyVarKind tv) == D env' (tyVarKind tv') && D (extendCME env tv) ty == D (extendCME env' tv') ty' (CoercionTy {}, CoercionTy {}) -> True _ -> False instance Outputable a => Outputable (TypeMapG a) where ppr m = text "TypeMap elts" <+> ppr (foldTM (:) m []) emptyT :: TypeMapX a emptyT = TM { tm_var = emptyTM , tm_app = EmptyMap , tm_tycon = emptyDNameEnv , tm_forall = EmptyMap , tm_tylit = emptyTyLitMap , tm_coerce = Nothing } mapT :: (a->b) -> TypeMapX a -> TypeMapX b mapT f (TM { tm_var = tvar, tm_app = tapp, tm_tycon = ttycon , tm_forall = tforall, tm_tylit = tlit , tm_coerce = tcoerce }) = TM { tm_var = mapTM f tvar , tm_app = mapTM (mapTM f) tapp , tm_tycon = mapTM f ttycon , tm_forall = mapTM (mapTM f) tforall , tm_tylit = mapTM f tlit , tm_coerce = fmap f tcoerce } ----------------- lkT :: DeBruijn Type -> TypeMapX a -> Maybe a lkT (D env ty) m = go ty m where go ty | Just ty' <- trieMapView ty = go ty' go (TyVarTy v) = tm_var >.> lkVar env v go (AppTy t1 t2) = tm_app >.> lkG (D env t1) >=> lkG (D env t2) go (TyConApp tc []) = tm_tycon >.> lkDNamed tc go ty@(TyConApp _ (_:_)) = pprPanic "lkT TyConApp" (ppr ty) go (LitTy l) = tm_tylit >.> lkTyLit l go (ForAllTy (Named tv _) ty) = tm_forall >.> lkG (D (extendCME env tv) ty) >=> lkBndr env tv go ty@(ForAllTy (Anon _) _) = pprPanic "lkT FunTy" (ppr ty) go (CastTy t _) = go t go (CoercionTy {}) = tm_coerce ----------------- xtT :: DeBruijn Type -> XT a -> TypeMapX a -> TypeMapX a xtT (D env ty) f m | Just ty' <- trieMapView ty = xtT (D env ty') f m xtT (D env (TyVarTy v)) f m = m { tm_var = tm_var m |> xtVar env v f } xtT (D env (AppTy t1 t2)) f m = m { tm_app = tm_app m |> xtG (D env t1) |>> xtG (D env t2) f } xtT (D _ (TyConApp tc [])) f m = m { tm_tycon = tm_tycon m |> xtDNamed tc f } xtT (D _ (LitTy l)) f m = m { tm_tylit = tm_tylit m |> xtTyLit l f } xtT (D env (CastTy t _)) f m = xtT (D env t) f m xtT (D _ (CoercionTy {})) f m = m { tm_coerce = tm_coerce m |> f } xtT (D env (ForAllTy (Named tv _) ty)) f m = m { tm_forall = tm_forall m |> xtG (D (extendCME env tv) ty) |>> xtBndr env tv f } xtT (D _ ty@(TyConApp _ (_:_))) _ _ = pprPanic "xtT TyConApp" (ppr ty) xtT (D _ ty@(ForAllTy (Anon _) _)) _ _ = pprPanic "xtT FunTy" (ppr ty) fdT :: (a -> b -> b) -> TypeMapX a -> b -> b fdT k m = foldTM k (tm_var m) . foldTM (foldTM k) (tm_app m) . foldTM k (tm_tycon m) . foldTM (foldTM k) (tm_forall m) . foldTyLit k (tm_tylit m) . foldMaybe k (tm_coerce m) ------------------------ data TyLitMap a = TLM { tlm_number :: Map.Map Integer a , tlm_string :: Map.Map FastString a } instance TrieMap TyLitMap where type Key TyLitMap = TyLit emptyTM = emptyTyLitMap lookupTM = lkTyLit alterTM = xtTyLit foldTM = foldTyLit mapTM = mapTyLit emptyTyLitMap :: TyLitMap a emptyTyLitMap = TLM { tlm_number = Map.empty, tlm_string = Map.empty } mapTyLit :: (a->b) -> TyLitMap a -> TyLitMap b mapTyLit f (TLM { tlm_number = tn, tlm_string = ts }) = TLM { tlm_number = Map.map f tn, tlm_string = Map.map f ts } lkTyLit :: TyLit -> TyLitMap a -> Maybe a lkTyLit l = case l of NumTyLit n -> tlm_number >.> Map.lookup n StrTyLit n -> tlm_string >.> Map.lookup n xtTyLit :: TyLit -> XT a -> TyLitMap a -> TyLitMap a xtTyLit l f m = case l of NumTyLit n -> m { tlm_number = tlm_number m |> Map.alter f n } StrTyLit n -> m { tlm_string = tlm_string m |> Map.alter f n } foldTyLit :: (a -> b -> b) -> TyLitMap a -> b -> b foldTyLit l m = flip (Map.fold l) (tlm_string m) . flip (Map.fold l) (tlm_number m) ------------------------------------------------- -- | @TypeMap a@ is a map from 'Type' to @a@. If you are a client, this -- is the type you want. The keys in this map may have different kinds. newtype TypeMap a = TypeMap (TypeMapG (TypeMapG a)) lkTT :: DeBruijn Type -> TypeMap a -> Maybe a lkTT (D env ty) (TypeMap m) = lkG (D env $ typeKind ty) m >>= lkG (D env ty) xtTT :: DeBruijn Type -> XT a -> TypeMap a -> TypeMap a xtTT (D env ty) f (TypeMap m) = TypeMap (m |> xtG (D env $ typeKind ty) |>> xtG (D env ty) f) -- Below are some client-oriented functions which operate on 'TypeMap'. instance TrieMap TypeMap where type Key TypeMap = Type emptyTM = TypeMap emptyTM lookupTM k m = lkTT (deBruijnize k) m alterTM k f m = xtTT (deBruijnize k) f m foldTM k (TypeMap m) = foldTM (foldTM k) m mapTM f (TypeMap m) = TypeMap (mapTM (mapTM f) m) foldTypeMap :: (a -> b -> b) -> b -> TypeMap a -> b foldTypeMap k z m = foldTM k m z emptyTypeMap :: TypeMap a emptyTypeMap = emptyTM lookupTypeMap :: TypeMap a -> Type -> Maybe a lookupTypeMap cm t = lookupTM t cm extendTypeMap :: TypeMap a -> Type -> a -> TypeMap a extendTypeMap m t v = alterTM t (const (Just v)) m -- | A 'LooseTypeMap' doesn't do a kind-check. Thus, when lookup up (t |> g), -- you'll find entries inserted under (t), even if (g) is non-reflexive. newtype LooseTypeMap a = LooseTypeMap (TypeMapG a) instance TrieMap LooseTypeMap where type Key LooseTypeMap = Type emptyTM = LooseTypeMap emptyTM lookupTM k (LooseTypeMap m) = lookupTM (deBruijnize k) m alterTM k f (LooseTypeMap m) = LooseTypeMap (alterTM (deBruijnize k) f m) foldTM f (LooseTypeMap m) = foldTM f m mapTM f (LooseTypeMap m) = LooseTypeMap (mapTM f m) {- ************************************************************************ * * Variables * * ************************************************************************ -} type BoundVar = Int -- Bound variables are deBruijn numbered type BoundVarMap a = IntMap.IntMap a data CmEnv = CME { cme_next :: BoundVar , cme_env :: VarEnv BoundVar } emptyCME :: CmEnv emptyCME = CME { cme_next = 0, cme_env = emptyVarEnv } extendCME :: CmEnv -> Var -> CmEnv extendCME (CME { cme_next = bv, cme_env = env }) v = CME { cme_next = bv+1, cme_env = extendVarEnv env v bv } extendCMEs :: CmEnv -> [Var] -> CmEnv extendCMEs env vs = foldl extendCME env vs lookupCME :: CmEnv -> Var -> Maybe BoundVar lookupCME (CME { cme_env = env }) v = lookupVarEnv env v -- | @DeBruijn a@ represents @a@ modulo alpha-renaming. This is achieved -- by equipping the value with a 'CmEnv', which tracks an on-the-fly deBruijn -- numbering. This allows us to define an 'Eq' instance for @DeBruijn a@, even -- if this was not (easily) possible for @a@. Note: we purposely don't -- export the constructor. Make a helper function if you find yourself -- needing it. data DeBruijn a = D CmEnv a -- | Synthesizes a @DeBruijn a@ from an @a@, by assuming that there are no -- bound binders (an empty 'CmEnv'). This is usually what you want if there -- isn't already a 'CmEnv' in scope. deBruijnize :: a -> DeBruijn a deBruijnize = D emptyCME instance Eq (DeBruijn a) => Eq (DeBruijn [a]) where D _ [] == D _ [] = True D env (x:xs) == D env' (x':xs') = D env x == D env' x' && D env xs == D env' xs' _ == _ = False --------- Variable binders ------------- -- | A 'BndrMap' is a 'TypeMapG' which allows us to distinguish between -- binding forms whose binders have different types. For example, -- if we are doing a 'TrieMap' lookup on @\(x :: Int) -> ()@, we should -- not pick up an entry in the 'TrieMap' for @\(x :: Bool) -> ()@: -- we can disambiguate this by matching on the type (or kind, if this -- a binder in a type) of the binder. type BndrMap = TypeMapG -- Note [Binders] -- ~~~~~~~~~~~~~~ -- We need to use 'BndrMap' for 'Coercion', 'CoreExpr' AND 'Type', since all -- of these data types have binding forms. lkBndr :: CmEnv -> Var -> BndrMap a -> Maybe a lkBndr env v m = lkG (D env (varType v)) m xtBndr :: CmEnv -> Var -> XT a -> BndrMap a -> BndrMap a xtBndr env v f = xtG (D env (varType v)) f --------- Variable occurrence ------------- data VarMap a = VM { vm_bvar :: BoundVarMap a -- Bound variable , vm_fvar :: DVarEnv a } -- Free variable instance TrieMap VarMap where type Key VarMap = Var emptyTM = VM { vm_bvar = IntMap.empty, vm_fvar = emptyDVarEnv } lookupTM = lkVar emptyCME alterTM = xtVar emptyCME foldTM = fdVar mapTM = mapVar mapVar :: (a->b) -> VarMap a -> VarMap b mapVar f (VM { vm_bvar = bv, vm_fvar = fv }) = VM { vm_bvar = mapTM f bv, vm_fvar = mapTM f fv } lkVar :: CmEnv -> Var -> VarMap a -> Maybe a lkVar env v | Just bv <- lookupCME env v = vm_bvar >.> lookupTM bv | otherwise = vm_fvar >.> lkDFreeVar v xtVar :: CmEnv -> Var -> XT a -> VarMap a -> VarMap a xtVar env v f m | Just bv <- lookupCME env v = m { vm_bvar = vm_bvar m |> alterTM bv f } | otherwise = m { vm_fvar = vm_fvar m |> xtDFreeVar v f } fdVar :: (a -> b -> b) -> VarMap a -> b -> b fdVar k m = foldTM k (vm_bvar m) . foldTM k (vm_fvar m) lkDFreeVar :: Var -> DVarEnv a -> Maybe a lkDFreeVar var env = lookupDVarEnv env var xtDFreeVar :: Var -> XT a -> DVarEnv a -> DVarEnv a xtDFreeVar v f m = alterDVarEnv f m v
vikraman/ghc
compiler/coreSyn/TrieMap.hs
bsd-3-clause
41,908
0
18
13,468
12,584
6,374
6,210
652
11
{-# LANGUAGE CPP #-} {-# LANGUAGE DeriveDataTypeable #-} {-# LANGUAGE OverloadedStrings #-} {-# LANGUAGE ScopedTypeVariables #-} {-# LANGUAGE FlexibleContexts #-} {-# LANGUAGE TemplateHaskell #-} {-# LANGUAGE RecordWildCards #-} {-# LANGUAGE ViewPatterns #-} -- | Main stack tool entry point. module Main (main) where import Control.Exception import qualified Control.Exception.Lifted as EL import Control.Monad hiding (mapM, forM) import Control.Monad.Catch (MonadThrow) import Control.Monad.IO.Class import Control.Monad.Logger import Control.Monad.Reader (ask, asks, runReaderT) import Control.Monad.Trans.Control (MonadBaseControl) import Control.Monad.Trans.Either (EitherT) import Control.Monad.Trans.Writer (Writer) import Data.Attoparsec.Args (getInterpreterArgs, parseArgs, EscapingMode (Escaping)) import qualified Data.ByteString.Lazy as L import Data.IORef import Data.List import qualified Data.Map as Map import qualified Data.Map.Strict as M import Data.Maybe import Data.Maybe.Extra (mapMaybeA) import Data.Monoid import qualified Data.Set as Set import Data.Text (Text) import qualified Data.Text as T import qualified Data.Text.IO as T import Data.Traversable import Data.Typeable (Typeable) import Data.Version (showVersion) #ifdef USE_GIT_INFO import Development.GitRev (gitCommitCount, gitHash) #endif import Distribution.System (buildArch, buildPlatform) import Distribution.Text (display) import GHC.IO.Encoding (mkTextEncoding, textEncodingName) import Network.HTTP.Client import Options.Applicative import Options.Applicative.Args import Options.Applicative.Help(footerHelp,stringChunk) import Options.Applicative.Builder.Extra import Options.Applicative.Complicated #ifdef USE_GIT_INFO import Options.Applicative.Simple (simpleVersion) #endif import Options.Applicative.Types (readerAsk, ParserHelp) import Path import Path.Extra (toFilePathNoTrailingSep) import Path.IO import qualified Paths_stack as Meta import Prelude hiding (pi, mapM) import Stack.Build import Stack.Clean (CleanOpts, clean) import Stack.Config import Stack.ConfigCmd as ConfigCmd import Stack.Constants import Stack.Coverage import qualified Stack.Docker as Docker import Stack.Dot import Stack.Exec import qualified Stack.Nix as Nix import Stack.Fetch import Stack.FileWatch import Stack.GhcPkg (getGlobalDB, mkGhcPackagePath) import Stack.Ghci import Stack.Ide import qualified Stack.Image as Image import Stack.Init import Stack.New import Stack.Options import Stack.Package (getCabalFileName) import qualified Stack.PackageIndex import Stack.SDist (getSDistTarball, checkSDistTarball, checkSDistTarball') import Stack.Setup import qualified Stack.Sig as Sig import Stack.Solver (solveExtraDeps) import Stack.Types import Stack.Types.Internal import Stack.Types.StackT import Stack.Upgrade import qualified Stack.Upload as Upload import System.Directory (canonicalizePath, doesFileExist, doesDirectoryExist, createDirectoryIfMissing) import qualified System.Directory as Directory (findExecutable) import System.Environment (getEnvironment, getProgName, getArgs, withArgs) import System.Exit import System.FileLock (lockFile, tryLockFile, unlockFile, SharedExclusive(Exclusive), FileLock) import System.FilePath (searchPathSeparator) import System.IO (hIsTerminalDevice, stderr, stdin, stdout, hSetBuffering, BufferMode(..), hPutStrLn, Handle, hGetEncoding, hSetEncoding) import System.Process.Read -- | Change the character encoding of the given Handle to transliterate -- on unsupported characters instead of throwing an exception hSetTranslit :: Handle -> IO () hSetTranslit h = do menc <- hGetEncoding h case fmap textEncodingName menc of Just name | '/' `notElem` name -> do enc' <- mkTextEncoding $ name ++ "//TRANSLIT" hSetEncoding h enc' _ -> return () versionString' :: String #ifdef USE_GIT_INFO versionString' = concat $ concat [ [$(simpleVersion Meta.version)] -- Leave out number of commits for --depth=1 clone -- See https://github.com/commercialhaskell/stack/issues/792 , [" (" ++ commitCount ++ " commits)" | commitCount /= ("1"::String) && commitCount /= ("UNKNOWN" :: String)] , [" ", display buildArch] ] where commitCount = $gitCommitCount #else versionString' = showVersion Meta.version ++ ' ' : display buildArch #endif main :: IO () main = do -- Line buffer the output by default, particularly for non-terminal runs. -- See https://github.com/commercialhaskell/stack/pull/360 hSetBuffering stdout LineBuffering hSetBuffering stdin LineBuffering hSetBuffering stderr LineBuffering hSetTranslit stdout hSetTranslit stderr args <- getArgs progName <- getProgName isTerminal <- hIsTerminalDevice stdout execExtraHelp args Docker.dockerHelpOptName (dockerOptsParser False) ("Only showing --" ++ Docker.dockerCmdName ++ "* options.") execExtraHelp args Nix.nixHelpOptName (nixOptsParser False) ("Only showing --" ++ Nix.nixCmdName ++ "* options.") eGlobalRun <- try $ commandLineHandler progName False case eGlobalRun of Left (exitCode :: ExitCode) -> do throwIO exitCode Right (globalMonoid,run) -> do let global = globalOptsFromMonoid isTerminal globalMonoid when (globalLogLevel global == LevelDebug) $ hPutStrLn stderr versionString' case globalReExecVersion global of Just expectVersion | expectVersion /= showVersion Meta.version -> throwIO $ InvalidReExecVersion expectVersion (showVersion Meta.version) _ -> return () run global `catch` \e -> -- This special handler stops "stack: " from being printed before the -- exception case fromException e of Just ec -> exitWith ec Nothing -> do printExceptionStderr e exitFailure commandLineHandler :: String -> Bool -> IO (GlobalOptsMonoid, GlobalOpts -> IO ()) commandLineHandler progName isInterpreter = complicatedOptions Meta.version (Just versionString') "stack - The Haskell Tool Stack" "" (globalOpts False) (Just failureCallback) (addCommands (globalOpts True) isInterpreter) where failureCallback f args = case stripPrefix "Invalid argument" (fst (renderFailure f "")) of Just _ -> if isInterpreter then handleParseResult (Failure f) else secondaryCommandHandler args >>= maybe (interpreterHandler f args) id Nothing -> handleParseResult (Failure f) globalOpts hide = extraHelpOption hide progName (Docker.dockerCmdName ++ "*") Docker.dockerHelpOptName <*> extraHelpOption hide progName (Nix.nixCmdName ++ "*") Nix.nixHelpOptName <*> globalOptsParser hide (if isInterpreter then Just $ LevelOther "silent" else Nothing) globalFooter :: String globalFooter = "Run 'stack --help' for global options that apply to all subcommands." addCommands :: Monoid c => Parser c -> Bool -> EitherT (GlobalOpts -> IO ()) (Writer (Mod CommandFields (GlobalOpts -> IO (), c))) () addCommands globalOpts isInterpreter = do when (not isInterpreter) (do addCommand' "build" "Build the package(s) in this directory/configuration" buildCmd (buildOptsParser Build) addCommand' "install" "Shortcut for 'build --copy-bins'" buildCmd (buildOptsParser Install) addCommand' "uninstall" "DEPRECATED: This command performs no actions, and is present for documentation only" uninstallCmd (many $ strArgument $ metavar "IGNORED") addCommand' "test" "Shortcut for 'build --test'" buildCmd (buildOptsParser Test) addCommand' "bench" "Shortcut for 'build --bench'" buildCmd (buildOptsParser Bench) addCommand' "haddock" "Shortcut for 'build --haddock'" buildCmd (buildOptsParser Haddock) addCommand' "new" "Create a new project from a template. Run `stack templates' to see available templates." newCmd newOptsParser addCommand' "templates" "List the templates available for `stack new'." templatesCmd (pure ()) addCommand' "init" "Initialize a stack project based on one or more cabal packages" initCmd initOptsParser addCommand' "solver" "Use a dependency solver to try and determine missing extra-deps" solverCmd solverOptsParser addCommand' "setup" "Get the appropriate GHC for your project" setupCmd setupParser addCommand' "path" "Print out handy path information" pathCmd (mapMaybeA (\(desc,name,_) -> flag Nothing (Just name) (long (T.unpack name) <> help desc)) paths) addCommand' "unpack" "Unpack one or more packages locally" unpackCmd (some $ strArgument $ metavar "PACKAGE") addCommand' "update" "Update the package index" updateCmd (pure ()) addCommand' "upgrade" "Upgrade to the latest stack (experimental)" upgradeCmd ((,) <$> switch ( long "git" <> help "Clone from Git instead of downloading from Hackage (more dangerous)" ) <*> strOption ( long "git-repo" <> help "Clone from specified git repository" <> value "https://github.com/commercialhaskell/stack" <> showDefault )) addCommand' "upload" "Upload a package to Hackage" uploadCmd ((,,,) <$> many (strArgument $ metavar "TARBALL/DIR") <*> optional pvpBoundsOption <*> ignoreCheckSwitch <*> flag False True (long "sign" <> help "GPG sign & submit signature")) addCommand' "sdist" "Create source distribution tarballs" sdistCmd ((,,) <$> many (strArgument $ metavar "DIR") <*> optional pvpBoundsOption <*> ignoreCheckSwitch) addCommand' "dot" "Visualize your project's dependency graph using Graphviz dot" dotCmd dotOptsParser addCommand' "exec" "Execute a command" execCmd (execOptsParser Nothing) addCommand' "ghc" "Run ghc" execCmd (execOptsParser $ Just ExecGhc) addCommand' "ghci" "Run ghci in the context of package(s) (experimental)" ghciCmd ghciOptsParser addCommand' "repl" "Run ghci in the context of package(s) (experimental) (alias for 'ghci')" ghciCmd ghciOptsParser ) -- These two are the only commands allowed in interpreter mode as well addCommand' "runghc" "Run runghc" execCmd (execOptsParser $ Just ExecRunGhc) addCommand' "runhaskell" "Run runghc (alias for 'runghc')" execCmd (execOptsParser $ Just ExecRunGhc) when (not isInterpreter) (do addCommand' "eval" "Evaluate some haskell code inline. Shortcut for 'stack exec ghc -- -e CODE'" evalCmd (evalOptsParser "CODE") addCommand' "clean" "Clean the local packages" cleanCmd cleanOptsParser addCommand' "list-dependencies" "List the dependencies" listDependenciesCmd (textOption (long "separator" <> metavar "SEP" <> help ("Separator between package name " <> "and package version.") <> value " " <> showDefault)) addCommand' "query" "Query general build information (experimental)" queryCmd (many $ strArgument $ metavar "SELECTOR...") addSubCommands' "ide" "IDE-specific commands" (do addCommand' "start" "Start the ide-backend service" ideCmd ((,) <$> many (textArgument (metavar "TARGET" <> help ("If none specified, use all " <> "packages defined in current directory"))) <*> argsOption (long "ghc-options" <> metavar "OPTION" <> help "Additional options passed to GHCi" <> value [])) addCommand' "packages" "List all available local loadable packages" packagesCmd (pure ()) addCommand' "load-targets" "List all load targets for a package target" targetsCmd (textArgument (metavar "TARGET"))) addSubCommands' Docker.dockerCmdName "Subcommands specific to Docker use" (do addCommand' Docker.dockerPullCmdName "Pull latest version of Docker image from registry" dockerPullCmd (pure ()) addCommand' "reset" "Reset the Docker sandbox" dockerResetCmd (switch (long "keep-home" <> help "Do not delete sandbox's home directory")) addCommand' Docker.dockerCleanupCmdName "Clean up Docker images and containers" dockerCleanupCmd dockerCleanupOptsParser) addSubCommands' ConfigCmd.cfgCmdName "Subcommands specific to modifying stack.yaml files" (addCommand' ConfigCmd.cfgCmdSetName "Sets a field in the project's stack.yaml to value" cfgSetCmd configCmdSetParser) addSubCommands' Image.imgCmdName "Subcommands specific to imaging (EXPERIMENTAL)" (addCommand' Image.imgDockerCmdName "Build a Docker image for the project" imgDockerCmd (boolFlags True "build" "building the project before creating the container" idm)) addSubCommands' "hpc" "Subcommands specific to Haskell Program Coverage" (addCommand' "report" "Generate HPC report a combined HPC report" hpcReportCmd hpcReportOptsParser) addSubCommands' Sig.sigCmdName "Subcommands specific to package signatures (EXPERIMENTAL)" (addSubCommands' Sig.sigSignCmdName "Sign a a single package or all your packages" (addCommand' Sig.sigSignSdistCmdName "Sign a single sdist package file" sigSignSdistCmd Sig.sigSignSdistOpts)) ) where ignoreCheckSwitch = switch (long "ignore-check" <> help "Do not check package for common mistakes") -- addCommand hiding global options addCommand' cmd title constr = addCommand cmd title globalFooter constr globalOpts addSubCommands' cmd title = addSubCommands cmd title globalFooter globalOpts secondaryCommandHandler :: (MonadIO m, MonadThrow m, MonadBaseControl IO m) => [String] -> IO (Maybe (m a)) -- fall-through to external executables in `git` style if they exist -- (i.e. `stack something` looks for `stack-something` before -- failing with "Invalid argument `something'") secondaryCommandHandler args = do -- FIXME this is broken when any options are specified before the command -- e.g. stack --verbosity silent cmd mExternalExec <- Directory.findExecutable ("stack-" ++ head args) case mExternalExec of Just ex -> do menv <- getEnvOverride buildPlatform return (Just $ runNoLoggingT (exec menv ex (tail args))) Nothing -> return Nothing interpreterHandler :: Monoid t => ParserFailure ParserHelp -> [String] -> IO (GlobalOptsMonoid, (GlobalOpts -> IO (), t)) interpreterHandler f args = do val <- getInterpreterArgs args stackProgName case val of Nothing -> do let hlp = footerHelp $ stringChunk $ concat [ "\nIf you are trying to use " , stackProgName , " as a script interpreter, a\n'-- " , stackProgName , " [options] runghc [options]' comment is required." , "\nSee https://github.com/commercialhaskell/stack/blob/release/doc/GUIDE.md#ghcrunghc" ] handleParseResult (overFailure (mappend hlp) (Failure f)) Just iargs -> do progName <- getProgName let cmdlineParse = commandLineHandler progName True (a,b) <- withArgs (iargs ++ "--" : args) cmdlineParse return (a,(b,mempty)) -- | Print out useful path information in a human-readable format (and -- support others later). pathCmd :: [Text] -> GlobalOpts -> IO () pathCmd keys go = withBuildConfig go (do env <- ask let cfg = envConfig env bc = envConfigBuildConfig cfg -- This is the modified 'bin-path', -- including the local GHC or MSYS if not configured to operate on -- global GHC. -- It was set up in 'withBuildConfigAndLock -> withBuildConfigExt -> setupEnv'. -- So it's not the *minimal* override path. menv <- getMinimalEnvOverride snap <- packageDatabaseDeps local <- packageDatabaseLocal extra <- packageDatabaseExtra global <- getGlobalDB menv =<< getWhichCompiler snaproot <- installationRootDeps localroot <- installationRootLocal distDir <- distRelativeDir hpcDir <- hpcReportDir forM_ -- filter the chosen paths in flags (keys), -- or show all of them if no specific paths chosen. (filter (\(_,key,_) -> null keys || elem key keys) paths) (\(_,key,path) -> liftIO $ T.putStrLn -- If a single path type is requested, output it directly. -- Otherwise, name all the paths. ((if length keys == 1 then "" else key <> ": ") <> path (PathInfo bc menv snap local global snaproot localroot distDir hpcDir extra)))) -- | Passed to all the path printers as a source of info. data PathInfo = PathInfo { piBuildConfig :: BuildConfig , piEnvOverride :: EnvOverride , piSnapDb :: Path Abs Dir , piLocalDb :: Path Abs Dir , piGlobalDb :: Path Abs Dir , piSnapRoot :: Path Abs Dir , piLocalRoot :: Path Abs Dir , piDistDir :: Path Rel Dir , piHpcDir :: Path Abs Dir , piExtraDbs :: [Path Abs Dir] } -- | The paths of interest to a user. The first tuple string is used -- for a description that the optparse flag uses, and the second -- string as a machine-readable key and also for @--foo@ flags. The user -- can choose a specific path to list like @--global-stack-root@. But -- really it's mainly for the documentation aspect. -- -- When printing output we generate @PathInfo@ and pass it to the -- function to generate an appropriate string. Trailing slashes are -- removed, see #506 paths :: [(String, Text, PathInfo -> Text)] paths = [ ( "Global stack root directory" , "global-stack-root" , T.pack . toFilePathNoTrailingSep . configStackRoot . bcConfig . piBuildConfig ) , ( "Project root (derived from stack.yaml file)" , "project-root" , T.pack . toFilePathNoTrailingSep . bcRoot . piBuildConfig ) , ( "Configuration location (where the stack.yaml file is)" , "config-location" , T.pack . toFilePath . bcStackYaml . piBuildConfig ) , ( "PATH environment variable" , "bin-path" , T.pack . intercalate [searchPathSeparator] . eoPath . piEnvOverride ) , ( "Installed GHCs (unpacked and archives)" , "ghc-paths" , T.pack . toFilePathNoTrailingSep . configLocalPrograms . bcConfig . piBuildConfig ) , ( "Local bin path where stack installs executables" , "local-bin-path" , T.pack . toFilePathNoTrailingSep . configLocalBin . bcConfig . piBuildConfig ) , ( "Extra include directories" , "extra-include-dirs" , T.intercalate ", " . Set.elems . configExtraIncludeDirs . bcConfig . piBuildConfig ) , ( "Extra library directories" , "extra-library-dirs" , T.intercalate ", " . Set.elems . configExtraLibDirs . bcConfig . piBuildConfig ) , ( "Snapshot package database" , "snapshot-pkg-db" , T.pack . toFilePathNoTrailingSep . piSnapDb ) , ( "Local project package database" , "local-pkg-db" , T.pack . toFilePathNoTrailingSep . piLocalDb ) , ( "Global package database" , "global-pkg-db" , T.pack . toFilePathNoTrailingSep . piGlobalDb ) , ( "GHC_PACKAGE_PATH environment variable" , "ghc-package-path" , \pi -> mkGhcPackagePath True (piLocalDb pi) (piSnapDb pi) (piExtraDbs pi) (piGlobalDb pi)) , ( "Snapshot installation root" , "snapshot-install-root" , T.pack . toFilePathNoTrailingSep . piSnapRoot ) , ( "Local project installation root" , "local-install-root" , T.pack . toFilePathNoTrailingSep . piLocalRoot ) , ( "Snapshot documentation root" , "snapshot-doc-root" , \pi -> T.pack (toFilePathNoTrailingSep (piSnapRoot pi </> docDirSuffix))) , ( "Local project documentation root" , "local-doc-root" , \pi -> T.pack (toFilePathNoTrailingSep (piLocalRoot pi </> docDirSuffix))) , ( "Dist work directory" , "dist-dir" , T.pack . toFilePathNoTrailingSep . piDistDir ) , ( "Where HPC reports and tix files are stored" , "local-hpc-root" , T.pack . toFilePathNoTrailingSep . piHpcDir ) ] data SetupCmdOpts = SetupCmdOpts { scoCompilerVersion :: !(Maybe CompilerVersion) , scoForceReinstall :: !Bool , scoUpgradeCabal :: !Bool , scoStackSetupYaml :: !String , scoGHCBindistURL :: !(Maybe String) } setupParser :: Parser SetupCmdOpts setupParser = SetupCmdOpts <$> optional (argument readVersion (metavar "GHC_VERSION" <> help ("Version of GHC to install, e.g. 7.10.2. " ++ "The default is to install the version implied by the resolver."))) <*> boolFlags False "reinstall" "reinstalling GHC, even if available (implies no-system-ghc)" idm <*> boolFlags False "upgrade-cabal" "installing the newest version of the Cabal library globally" idm <*> strOption ( long "stack-setup-yaml" <> help "Location of the main stack-setup.yaml file" <> value defaultStackSetupYaml <> showDefault ) <*> optional (strOption (long "ghc-bindist" <> metavar "URL" <> help "Alternate GHC binary distribution (requires custom --ghc-variant)")) where readVersion = do s <- readerAsk case parseCompilerVersion ("ghc-" <> T.pack s) of Nothing -> case parseCompilerVersion (T.pack s) of Nothing -> readerError $ "Invalid version: " ++ s Just x -> return x Just x -> return x setupCmd :: SetupCmdOpts -> GlobalOpts -> IO () setupCmd SetupCmdOpts{..} go@GlobalOpts{..} = do (manager,lc) <- loadConfigWithOpts go withUserFileLock go (configStackRoot $ lcConfig lc) $ \lk -> runStackTGlobal manager (lcConfig lc) go $ Docker.reexecWithOptionalContainer (lcProjectRoot lc) Nothing (runStackTGlobal manager (lcConfig lc) go $ Nix.reexecWithOptionalShell $ runStackLoggingTGlobal manager go $ do (wantedCompiler, compilerCheck, mstack) <- case scoCompilerVersion of Just v -> return (v, MatchMinor, Nothing) Nothing -> do bc <- lcLoadBuildConfig lc globalCompiler return ( bcWantedCompiler bc , configCompilerCheck (lcConfig lc) , Just $ bcStackYaml bc ) miniConfig <- loadMiniConfig (lcConfig lc) mpaths <- runStackTGlobal manager miniConfig go $ ensureCompiler SetupOpts { soptsInstallIfMissing = True , soptsUseSystem = configSystemGHC (lcConfig lc) && not scoForceReinstall , soptsWantedCompiler = wantedCompiler , soptsCompilerCheck = compilerCheck , soptsStackYaml = mstack , soptsForceReinstall = scoForceReinstall , soptsSanityCheck = True , soptsSkipGhcCheck = False , soptsSkipMsys = configSkipMsys $ lcConfig lc , soptsUpgradeCabal = scoUpgradeCabal , soptsResolveMissingGHC = Nothing , soptsStackSetupYaml = scoStackSetupYaml , soptsGHCBindistURL = scoGHCBindistURL } let compiler = case wantedCompiler of GhcVersion _ -> "GHC" GhcjsVersion {} -> "GHCJS" case mpaths of Nothing -> $logInfo $ "stack will use the " <> compiler <> " on your PATH" Just _ -> $logInfo $ "stack will use a locally installed " <> compiler $logInfo "For more information on paths, see 'stack path' and 'stack exec env'" $logInfo $ "To use this " <> compiler <> " and packages outside of a project, consider using:" $logInfo "stack ghc, stack ghci, stack runghc, or stack exec" ) Nothing (Just $ munlockFile lk) -- | Unlock a lock file, if the value is Just munlockFile :: MonadIO m => Maybe FileLock -> m () munlockFile Nothing = return () munlockFile (Just lk) = liftIO $ unlockFile lk -- | Enforce mutual exclusion of every action running via this -- function, on this path, on this users account. -- -- A lock file is created inside the given directory. Currently, -- stack uses locks per-snapshot. In the future, stack may refine -- this to an even more fine-grain locking approach. -- withUserFileLock :: (MonadBaseControl IO m, MonadIO m) => GlobalOpts -> Path Abs Dir -> (Maybe FileLock -> m a) -> m a withUserFileLock go@GlobalOpts{} dir act = do env <- liftIO getEnvironment let toLock = lookup "STACK_LOCK" env == Just "true" if toLock then do let lockfile = $(mkRelFile "lockfile") let pth = dir </> lockfile liftIO $ createDirectoryIfMissing True (toFilePath dir) -- Just in case of asynchronous exceptions, we need to be careful -- when using tryLockFile here: EL.bracket (liftIO $ tryLockFile (toFilePath pth) Exclusive) (maybe (return ()) (liftIO . unlockFile)) (\fstTry -> case fstTry of Just lk -> EL.finally (act $ Just lk) (liftIO $ unlockFile lk) Nothing -> do let chatter = globalLogLevel go /= LevelOther "silent" when chatter $ liftIO $ hPutStrLn stderr $ "Failed to grab lock ("++show pth++ "); other stack instance running. Waiting..." EL.bracket (liftIO $ lockFile (toFilePath pth) Exclusive) (liftIO . unlockFile) (\lk -> do when chatter $ liftIO $ hPutStrLn stderr "Lock acquired, proceeding." act $ Just lk)) else act Nothing withConfigAndLock :: GlobalOpts -> StackT Config IO () -> IO () withConfigAndLock go@GlobalOpts{..} inner = do (manager, lc) <- loadConfigWithOpts go withUserFileLock go (configStackRoot $ lcConfig lc) $ \lk -> runStackTGlobal manager (lcConfig lc) go $ Docker.reexecWithOptionalContainer (lcProjectRoot lc) Nothing (runStackTGlobal manager (lcConfig lc) go inner) Nothing (Just $ munlockFile lk) -- For now the non-locking version just unlocks immediately. -- That is, there's still a serialization point. withBuildConfig :: GlobalOpts -> StackT EnvConfig IO () -> IO () withBuildConfig go inner = withBuildConfigAndLock go (\lk -> do munlockFile lk inner) withBuildConfigAndLock :: GlobalOpts -> (Maybe FileLock -> StackT EnvConfig IO ()) -> IO () withBuildConfigAndLock go inner = withBuildConfigExt go Nothing inner Nothing withBuildConfigExt :: GlobalOpts -> Maybe (StackT Config IO ()) -- ^ Action to perform after before build. This will be run on the host -- OS even if Docker is enabled for builds. The build config is not -- available in this action, since that would require build tools to be -- installed on the host OS. -> (Maybe FileLock -> StackT EnvConfig IO ()) -- ^ Action that uses the build config. If Docker is enabled for builds, -- this will be run in a Docker container. -> Maybe (StackT Config IO ()) -- ^ Action to perform after the build. This will be run on the host -- OS even if Docker is enabled for builds. The build config is not -- available in this action, since that would require build tools to be -- installed on the host OS. -> IO () withBuildConfigExt go@GlobalOpts{..} mbefore inner mafter = do (manager, lc) <- loadConfigWithOpts go withUserFileLock go (configStackRoot $ lcConfig lc) $ \lk0 -> do -- A local bit of state for communication between callbacks: curLk <- newIORef lk0 let inner' lk = -- Locking policy: This is only used for build commands, which -- only need to lock the snapshot, not the global lock. We -- trade in the lock here. do dir <- installationRootDeps -- Hand-over-hand locking: withUserFileLock go dir $ \lk2 -> do liftIO $ writeIORef curLk lk2 liftIO $ munlockFile lk inner lk2 let inner'' lk = do bconfig <- runStackLoggingTGlobal manager go $ lcLoadBuildConfig lc globalCompiler envConfig <- runStackTGlobal manager bconfig go (setupEnv Nothing) runStackTGlobal manager envConfig go (inner' lk) runStackTGlobal manager (lcConfig lc) go $ Docker.reexecWithOptionalContainer (lcProjectRoot lc) mbefore (runStackTGlobal manager (lcConfig lc) go $ Nix.reexecWithOptionalShell (inner'' lk0)) mafter (Just $ liftIO $ do lk' <- readIORef curLk munlockFile lk') cleanCmd :: CleanOpts -> GlobalOpts -> IO () cleanCmd opts go = withBuildConfigAndLock go (const (clean opts)) -- | Helper for build and install commands buildCmd :: BuildOpts -> GlobalOpts -> IO () buildCmd opts go = do when (any (("-prof" `elem`) . either (const []) id . parseArgs Escaping) (boptsGhcOptions opts)) $ do hPutStrLn stderr "When building with stack, you should not use the -prof GHC option" hPutStrLn stderr "Instead, please use --library-profiling and --executable-profiling" hPutStrLn stderr "See: https://github.com/commercialhaskell/stack/issues/1015" error "-prof GHC option submitted" case boptsFileWatch opts of FileWatchPoll -> fileWatchPoll inner FileWatch -> fileWatch inner NoFileWatch -> inner $ const $ return () where inner setLocalFiles = withBuildConfigAndLock go $ \lk -> Stack.Build.build setLocalFiles lk opts uninstallCmd :: [String] -> GlobalOpts -> IO () uninstallCmd _ go = withConfigAndLock go $ do $logError "stack does not manage installations in global locations" $logError "The only global mutation stack performs is executable copying" $logError "For the default executable destination, please run 'stack path --local-bin-path'" -- | Unpack packages to the filesystem unpackCmd :: [String] -> GlobalOpts -> IO () unpackCmd names go = withConfigAndLock go $ do menv <- getMinimalEnvOverride Stack.Fetch.unpackPackages menv "." names -- | Update the package index updateCmd :: () -> GlobalOpts -> IO () updateCmd () go = withConfigAndLock go $ getMinimalEnvOverride >>= Stack.PackageIndex.updateAllIndices upgradeCmd :: (Bool, String) -> GlobalOpts -> IO () upgradeCmd (fromGit, repo) go = withConfigAndLock go $ upgrade (if fromGit then Just repo else Nothing) (globalResolver go) #ifdef USE_GIT_INFO (find (/= "UNKNOWN") [$gitHash]) #else Nothing #endif -- | Upload to Hackage uploadCmd :: ([String], Maybe PvpBounds, Bool, Bool) -> GlobalOpts -> IO () uploadCmd ([], _, _, _) _ = error "To upload the current package, please run 'stack upload .'" uploadCmd (args, mpvpBounds, ignoreCheck, shouldSign) go = do let partitionM _ [] = return ([], []) partitionM f (x:xs) = do r <- f x (as, bs) <- partitionM f xs return $ if r then (x:as, bs) else (as, x:bs) (files, nonFiles) <- partitionM doesFileExist args (dirs, invalid) <- partitionM doesDirectoryExist nonFiles unless (null invalid) $ error $ "stack upload expects a list sdist tarballs or cabal directories. Can't find " ++ show invalid (_,lc) <- liftIO $ loadConfigWithOpts go let getUploader :: (HasStackRoot config, HasPlatform config, HasConfig config) => StackT config IO Upload.Uploader getUploader = do config <- asks getConfig manager <- asks envManager let uploadSettings = Upload.setGetManager (return manager) Upload.defaultUploadSettings liftIO $ Upload.mkUploader config uploadSettings sigServiceUrl = "https://sig.commercialhaskell.org/" withBuildConfigAndLock go $ \_ -> do uploader <- getUploader unless ignoreCheck $ mapM_ (parseRelAsAbsFile >=> checkSDistTarball) files forM_ files (\file -> do tarFile <- parseRelAsAbsFile file liftIO (Upload.upload uploader (toFilePath tarFile)) when shouldSign (Sig.sign (lcProjectRoot lc) sigServiceUrl tarFile)) unless (null dirs) $ forM_ dirs $ \dir -> do pkgDir <- parseRelAsAbsDir dir (tarName, tarBytes) <- getSDistTarball mpvpBounds pkgDir unless ignoreCheck $ checkSDistTarball' tarName tarBytes liftIO $ Upload.uploadBytes uploader tarName tarBytes tarPath <- parseRelFile tarName when shouldSign (Sig.signTarBytes (lcProjectRoot lc) sigServiceUrl tarPath tarBytes) sdistCmd :: ([String], Maybe PvpBounds, Bool) -> GlobalOpts -> IO () sdistCmd (dirs, mpvpBounds, ignoreCheck) go = withBuildConfig go $ do -- No locking needed. -- If no directories are specified, build all sdist tarballs. dirs' <- if null dirs then asks (Map.keys . envConfigPackages . getEnvConfig) else mapM (parseAbsDir <=< liftIO . canonicalizePath) dirs forM_ dirs' $ \dir -> do (tarName, tarBytes) <- getSDistTarball mpvpBounds dir distDir <- distDirFromDir dir tarPath <- (distDir </>) <$> parseRelFile tarName liftIO $ createTree $ parent tarPath liftIO $ L.writeFile (toFilePath tarPath) tarBytes unless ignoreCheck (checkSDistTarball tarPath) $logInfo $ "Wrote sdist tarball to " <> T.pack (toFilePath tarPath) -- | Execute a command. execCmd :: ExecOpts -> GlobalOpts -> IO () execCmd ExecOpts {..} go@GlobalOpts{..} = case eoExtra of ExecOptsPlain -> do (cmd, args) <- case (eoCmd, eoArgs) of (ExecCmd cmd, args) -> return (cmd, args) (ExecGhc, args) -> return ("ghc", args) (ExecRunGhc, args) -> return ("runghc", args) (manager,lc) <- liftIO $ loadConfigWithOpts go withUserFileLock go (configStackRoot $ lcConfig lc) $ \lk -> runStackTGlobal manager (lcConfig lc) go $ Docker.reexecWithOptionalContainer (lcProjectRoot lc) -- Unlock before transferring control away, whether using docker or not: (Just $ munlockFile lk) (runStackTGlobal manager (lcConfig lc) go $ do config <- asks getConfig menv <- liftIO $ configEnvOverride config plainEnvSettings Nix.reexecWithOptionalShell (runStackTGlobal manager (lcConfig lc) go $ exec menv cmd args)) Nothing Nothing -- Unlocked already above. ExecOptsEmbellished {..} -> withBuildConfigAndLock go $ \lk -> do config <- asks getConfig (cmd, args) <- case (eoCmd, eoArgs) of (ExecCmd cmd, args) -> return (cmd, args) (ExecGhc, args) -> execCompiler "" args -- NOTE: this won't currently work for GHCJS, because it doesn't have -- a runghcjs binary. It probably will someday, though. (ExecRunGhc, args) -> execCompiler "run" args let targets = concatMap words eoPackages unless (null targets) $ Stack.Build.build (const $ return ()) lk defaultBuildOpts { boptsTargets = map T.pack targets } munlockFile lk -- Unlock before transferring control away. menv <- liftIO $ configEnvOverride config eoEnvSettings exec menv cmd args where execCompiler cmdPrefix args = do wc <- getWhichCompiler let cmd = cmdPrefix ++ compilerExeName wc return (cmd, args) -- | Evaluate some haskell code inline. evalCmd :: EvalOpts -> GlobalOpts -> IO () evalCmd EvalOpts {..} go@GlobalOpts {..} = execCmd execOpts go where execOpts = ExecOpts { eoCmd = ExecGhc , eoArgs = ["-e", evalArg] , eoExtra = evalExtra } -- | Run GHCi in the context of a project. ghciCmd :: GhciOpts -> GlobalOpts -> IO () ghciCmd ghciOpts go@GlobalOpts{..} = withBuildConfigAndLock go $ \lk -> do let packageTargets = concatMap words (ghciAdditionalPackages ghciOpts) unless (null packageTargets) $ Stack.Build.build (const $ return ()) lk defaultBuildOpts { boptsTargets = map T.pack packageTargets } munlockFile lk -- Don't hold the lock while in the GHCI. ghci ghciOpts -- | Run ide-backend in the context of a project. ideCmd :: ([Text], [String]) -> GlobalOpts -> IO () ideCmd (targets,args) go@GlobalOpts{..} = withBuildConfig go $ -- No locking needed. ide targets args -- | List packages in the project. packagesCmd :: () -> GlobalOpts -> IO () packagesCmd () go@GlobalOpts{..} = withBuildConfig go $ do econfig <- asks getEnvConfig locals <- forM (M.toList (envConfigPackages econfig)) $ \(dir,_) -> do cabalfp <- getCabalFileName dir parsePackageNameFromFilePath cabalfp forM_ locals (liftIO . putStrLn . packageNameString) -- | List load targets for a package target. targetsCmd :: Text -> GlobalOpts -> IO () targetsCmd target go@GlobalOpts{..} = withBuildConfig go $ do let bopts = defaultBuildOpts { boptsTargets = [target] } (_realTargets,_,pkgs) <- ghciSetup bopts False False Nothing pwd <- getWorkingDir targets <- fmap (concat . snd . unzip) (mapM (getPackageOptsAndTargetFiles pwd) pkgs) forM_ targets (liftIO . putStrLn) -- | Pull the current Docker image. dockerPullCmd :: () -> GlobalOpts -> IO () dockerPullCmd _ go@GlobalOpts{..} = do (manager,lc) <- liftIO $ loadConfigWithOpts go -- TODO: can we eliminate this lock if it doesn't touch ~/.stack/? withUserFileLock go (configStackRoot $ lcConfig lc) $ \_ -> runStackTGlobal manager (lcConfig lc) go $ Docker.preventInContainer Docker.pull -- | Reset the Docker sandbox. dockerResetCmd :: Bool -> GlobalOpts -> IO () dockerResetCmd keepHome go@GlobalOpts{..} = do (manager,lc) <- liftIO (loadConfigWithOpts go) -- TODO: can we eliminate this lock if it doesn't touch ~/.stack/? withUserFileLock go (configStackRoot $ lcConfig lc) $ \_ -> runStackTGlobal manager (lcConfig lc) go $ Docker.preventInContainer $ Docker.reset (lcProjectRoot lc) keepHome -- | Cleanup Docker images and containers. dockerCleanupCmd :: Docker.CleanupOpts -> GlobalOpts -> IO () dockerCleanupCmd cleanupOpts go@GlobalOpts{..} = do (manager,lc) <- liftIO $ loadConfigWithOpts go -- TODO: can we eliminate this lock if it doesn't touch ~/.stack/? withUserFileLock go (configStackRoot $ lcConfig lc) $ \_ -> runStackTGlobal manager (lcConfig lc) go $ Docker.preventInContainer $ Docker.cleanup cleanupOpts cfgSetCmd :: ConfigCmd.ConfigCmdSet -> GlobalOpts -> IO () cfgSetCmd co go@GlobalOpts{..} = withBuildConfigAndLock go (\_ -> do env <- ask runReaderT (cfgCmdSet co) env) imgDockerCmd :: Bool -> GlobalOpts -> IO () imgDockerCmd rebuild go@GlobalOpts{..} = withBuildConfigExt go Nothing (\lk -> do when rebuild $ Stack.Build.build (const (return ())) lk defaultBuildOpts Image.stageContainerImageArtifacts) (Just Image.createContainerImageFromStage) sigSignSdistCmd :: (String, String) -> GlobalOpts -> IO () sigSignSdistCmd (url,path) go = withConfigAndLock go (do (manager,lc) <- liftIO (loadConfigWithOpts go) tarBall <- parseRelAsAbsFile path runStackTGlobal manager (lcConfig lc) go (Sig.sign (lcProjectRoot lc) url tarBall)) -- | Load the configuration with a manager. Convenience function used -- throughout this module. loadConfigWithOpts :: GlobalOpts -> IO (Manager,LoadConfig (StackLoggingT IO)) loadConfigWithOpts go@GlobalOpts{..} = do manager <- newTLSManager mstackYaml <- case globalStackYaml of Nothing -> return Nothing Just fp -> do path <- canonicalizePath fp >>= parseAbsFile return $ Just path lc <- runStackLoggingTGlobal manager go $ do lc <- loadConfig globalConfigMonoid mstackYaml globalResolver -- If we have been relaunched in a Docker container, perform in-container initialization -- (switch UID, etc.). We do this after first loading the configuration since it must -- happen ASAP but needs a configuration. case globalDockerEntrypoint of Just de -> Docker.entrypoint (lcConfig lc) de Nothing -> return () return lc return (manager,lc) -- | Project initialization initCmd :: InitOpts -> GlobalOpts -> IO () initCmd initOpts go = withConfigAndLock go $ do pwd <- getWorkingDir config <- asks getConfig miniConfig <- loadMiniConfig config runReaderT (initProject pwd initOpts) miniConfig -- | Create a project directory structure and initialize the stack config. newCmd :: (NewOpts,InitOpts) -> GlobalOpts -> IO () newCmd (newOpts,initOpts) go@GlobalOpts{..} = withConfigAndLock go $ do dir <- new newOpts config <- asks getConfig miniConfig <- loadMiniConfig config runReaderT (initProject dir initOpts) miniConfig -- | List the available templates. templatesCmd :: () -> GlobalOpts -> IO () templatesCmd _ go@GlobalOpts{..} = withConfigAndLock go listTemplates -- | Fix up extra-deps for a project solverCmd :: Bool -- ^ modify stack.yaml automatically? -> GlobalOpts -> IO () solverCmd fixStackYaml go = withBuildConfigAndLock go (\_ -> solveExtraDeps fixStackYaml) -- | Visualize dependencies dotCmd :: DotOpts -> GlobalOpts -> IO () dotCmd dotOpts go = withBuildConfigAndLock go (\_ -> dot dotOpts) -- | List the dependencies listDependenciesCmd :: Text -> GlobalOpts -> IO () listDependenciesCmd sep go = withBuildConfig go (listDependencies sep') where sep' = T.replace "\\t" "\t" (T.replace "\\n" "\n" sep) -- | Query build information queryCmd :: [String] -> GlobalOpts -> IO () queryCmd selectors go = withBuildConfig go $ queryBuildInfo $ map T.pack selectors -- | Generate a combined HPC report hpcReportCmd :: HpcReportOpts -> GlobalOpts -> IO () hpcReportCmd hropts go = withBuildConfig go $ generateHpcReportForTargets hropts data MainException = InvalidReExecVersion String String deriving (Typeable) instance Exception MainException instance Show MainException where show (InvalidReExecVersion expected actual) = concat [ "When re-executing '" , stackProgName , "' in a container, the incorrect version was found\nExpected: " , expected , "; found: " , actual]
rubik/stack
src/main/Main.hs
bsd-3-clause
49,174
0
26
16,693
10,017
5,083
4,934
-1
-1
{-# LANGUAGE TemplateHaskell #-} module Lopc.Enums where import Database.Persist.TH data LopcClassification = MajorLopc | MinorLopc | OtherLopc deriving (Show, Read, Eq, Ord) derivePersistField "LopcClassification"
hectorhon/autotrace2
app/Lopc/Enums.hs
bsd-3-clause
266
0
6
75
51
29
22
7
0
-- -- GPIOF4.hs --- GPIO peripheral for F4 series & compatible -- -- Copyright (C) 2014, Galois, Inc. -- All Rights Reserved. -- module Ivory.BSP.STM32.Peripheral.GPIOF4 ( module Ivory.BSP.STM32.Peripheral.GPIOF4.Peripheral , module Ivory.BSP.STM32.Peripheral.GPIOF4.Regs , module Ivory.BSP.STM32.Peripheral.GPIOF4.RegTypes ) where import Ivory.BSP.STM32.Peripheral.GPIOF4.Peripheral import Ivory.BSP.STM32.Peripheral.GPIOF4.Regs import Ivory.BSP.STM32.Peripheral.GPIOF4.RegTypes
GaloisInc/ivory-tower-stm32
ivory-bsp-stm32/src/Ivory/BSP/STM32/Peripheral/GPIOF4.hs
bsd-3-clause
491
0
5
53
74
59
15
7
0
-- -*- mode: haskell; -*- -- -- generate anagrams based on dictionary and repeated user input -- import Data.Char (toLower) import Data.List (sort) import System.Environment (getArgs) import System.IO (stdout, hFlush) import qualified Data.Map as M (Map, empty, lookup, alter) -- map of canonical representation to all anagrams that exist type AnagramMap = M.Map String [String] dictionaryFilename :: IO String dictionaryFilename = do arg <- getArgs if null arg then (return "/usr/share/dict/words") else (return $ head arg) readDictionary :: String -> IO [String] readDictionary fileName = do fmap lines $ readFile fileName canonicalWord :: String -> String canonicalWord = sort . (map toLower) canonicalDictionary :: [String] -> AnagramMap canonicalDictionary ws = foldr go M.empty ws where go word acc = M.alter (updateWord word) (canonicalWord word) acc updateWord w m = Just [w] `mappend` m generateAnagrams :: AnagramMap -> String -> String generateAnagrams h w = maybe failureMessage formatResponse anagrams where anagrams = M.lookup (canonicalWord w) h formatResponse ws = "\nAnagrams for '" ++ w ++ "':\n" ++ (formatAnagrams ws) ++ "\nEnter a word to find anagrams for: " formatAnagrams arr = unlines $ map (\str -> "\t - '" ++ str ++ "'") arr failureMessage = "\nNo anagrams for '" ++ w ++ "'\nEnter a word to find anagrams for: " printAnagrams :: AnagramMap -> String -> String printAnagrams dictionary input = unlines $ map (generateAnagrams dictionary) (lines input) main :: IO () main = do dict <- fmap canonicalDictionary $ dictionaryFilename >>= readDictionary putStr "Enter a word to find anagrams for: " >> hFlush stdout interact $ printAnagrams dict
wiggly/workshop
app/Anagram.hs
bsd-3-clause
1,810
0
12
400
507
264
243
31
2
-- | Internal module containing the store definitions. {-# LANGUAGE AllowAmbiguousTypes, OverloadedStrings #-} {-# OPTIONS_GHC -fno-warn-redundant-constraints #-} module React.Flux.Store ( ReactStoreRef(..) , StoreData(..) -- * Old Stores , SomeStoreAction(..) , executeAction -- * New Stores , NewReactStoreHS(..) , someStoreAction , registerInitialStore , transformStore , readStoreData -- * Util , typeJsKey ) where import Control.Concurrent.MVar (MVar, newMVar, modifyMVar_, readMVar) import Control.DeepSeq import Data.Typeable import GHC.Generics (Generic) import React.Flux.Internal (unsafeDerefExport) import Data.Monoid ((<>)) import GHCJS.Foreign.Export import GHCJS.Types (JSVal, IsJSVal, JSString) import GHC.Fingerprint.Type import qualified Data.JSString.Int as JSString (decimal) import qualified Data.JSString as JSString import Data.Word (Word64) -- | See https://github.com/ghcjs/ghcjs/issues/570 for details. decimal_workaround_570 :: Word64 -> JSString decimal_workaround_570 w = dropleadingzeros . mconcat $ showpadded <$> chunks where n :: Integer -> Integer n i = 10^(5 * i) chunks :: [Integer] chunks = [ (fromIntegral w `div` (n 3)) `mod` n 1 , (fromIntegral w `div` (n 2)) `mod` n 1 , (fromIntegral w `div` (n 1)) `mod` n 1 , fromIntegral w `mod` n 1 ] showpadded :: Integer -> JSString showpadded i = JSString.reverse . JSString.take 5 . JSString.reverse $ JSString.pack "00000" <> JSString.decimal i dropleadingzeros :: JSString -> JSString dropleadingzeros = JSString.dropWhile (== '0') -- | A store contains application state, receives actions from the dispatcher, and notifies -- controller-views to re-render themselves. You can have multiple stores; it should be the case -- that all of the state required to render the page is contained in the stores. A store keeps a -- global reference to a value of type @storeData@, which must be an instance of 'StoreData'. -- -- Stores also work when compiled with GHC instead of GHCJS. When compiled with GHC, the store is -- just an MVar containing the store data and there are no controller views. 'alterStore' can still -- be used, but it just 'transform's the store and does not notify any controller-views since there -- are none. Compiling with GHC instead of GHCJS can be helpful for unit testing, although GHCJS -- plus node can also be used for unit testing. -- -- >data Todo = Todo { -- > todoText :: Text -- > , todoComplete :: Bool -- > , todoIsEditing :: Bool -- >} deriving (Show, Typeable) -- > -- >newtype TodoState = TodoState { -- > todoList :: [(Int, Todo)] -- >} deriving (Show, Typeable) -- > -- >data TodoAction = TodoCreate Text -- > | TodoDelete Int -- > | TodoEdit Int -- > | UpdateText Int Text -- > | ToggleAllComplete -- > | TodoSetComplete Int Bool -- > | ClearCompletedTodos -- > deriving (Show, Typeable, Generic, NFData) -- > -- >instance StoreData TodoState where -- > type StoreAction TodoState = TodoAction -- > transform action (TodoState todos) = ... -- > -- >initTodoStore :: IO () -- >initTodoStore = registerInitialStore $ TodoState -- > [ (0, Todo "Learn react" True False) -- > , (1, Todo "Learn react-flux" False False) -- > ] class Typeable storeData => StoreData storeData where -- | The actions that this store accepts type StoreAction storeData -- | Transform the store data according to the action. This is the only place in your app where -- @IO@ should occur. The transform function should complete quickly, since the UI will not be -- re-rendered until the transform is complete. Therefore, if you need to perform some longer -- action, you should fork a thread from inside 'transform'. The thread can then call 'alterStore' -- with another action with the result of its computation. This is very common to communicate with -- the backend using AJAX. Indeed, the 'React.Flux.Combinators.jsonAjax' utility function -- implements exactly this strategy since it is so common. -- -- Note that if the transform throws an exception, the transform will be aborted and the old -- store data will be kept unchanged. The exception will then be thrown from 'alterStore'. -- -- For the best performance, care should be taken in only modifying the part of the store data -- that changed (see below for more information on performance). transform :: StoreAction storeData -> storeData -> IO storeData -- | An existential type for some store action. It is used as the output of the dispatcher. -- The 'NFData' instance is important for performance, for details see below. data SomeStoreAction = forall storeData. (StoreData storeData, NFData (StoreAction storeData)) => SomeNewStoreAction (Proxy storeData) (StoreAction storeData) instance NFData SomeStoreAction where rnf (SomeNewStoreAction _ action) = action `deepseq` () -- | Create some store action. You must use a type-argument to specify the storeData type (because technically, the same -- store action type could be used for different stores). I strongly suggest you keep a one-to-one correspondence between -- stores and store actions, but GHC does not know that. For example, -- -- >todoAction :: TodoAction -> SomeStoreAction -- >todoAction a = someStoreAction @TodoStore a someStoreAction :: forall storeData. (StoreData storeData, NFData (StoreAction storeData)) => StoreAction storeData -> SomeStoreAction someStoreAction = SomeNewStoreAction (Proxy :: Proxy storeData) -- | Call 'alterStore' on the store and action. executeAction :: SomeStoreAction -> IO () executeAction (SomeNewStoreAction p a) = transformStore p a -------------------------------------------------------------------------------- -- Old Version Store Definition -------------------------------------------------------------------------------- -- | This type is used to represent the foreign javascript object part of the store. newtype ReactStoreRef storeData = ReactStoreRef JSVal deriving (Generic) instance IsJSVal (ReactStoreRef storeData) data NewReactStoreHS = NewReactStoreHS { newStoreLock :: MVar () } deriving Typeable ---------------------------------------------------------------------------------------------------- -- Store operations ---------------------------------------------------------------------------------------------------- -- | The new type of stores, introduced in version 1.3, keep the data in a javascript dictionary indexed -- by the fingerprint of the type. This allows any code to lookup the store by knowing the type. newtype NewReactStore storeData = NewReactStore JSVal instance IsJSVal (NewReactStore storeData) -- | New stores are kept in a javascript dictionary by type. This computes the key into this dictionary. -- -- FIXME: make the return value of this a newtype StoreKey, make typeJsKey private, and make -- Views.stateForView use StoreKey as hashmap keys. storeJsKey :: Typeable ty => Proxy ty -> JSString storeJsKey p = typeJsKey (typeRep p) typeJsKey :: TypeRep -> JSString typeJsKey t = decimal_workaround_570 f1 <> "-" <> decimal_workaround_570 f2 where Fingerprint f1 f2 = typeRepFingerprint t -- | Register the initial store data. This function must be called exactly once from your main function before -- the initial rendering occurs. Store data is global and so there can be only one store data value for each -- store type. -- -- FIXME: why not @storeE <- export =<< newMVar initial@? i think the reason to keep the lock -- separately in 'NewReactStoreHS' is just due to the fact that too much code is on the javascript -- side. The MVar should contain the entirety of @{ sdata: ..., views: ..., hs: <lock> }@, which -- would eliminate the need fo the lock. (medium-sized refactoring if we're lucky.) registerInitialStore :: forall storeData. (Typeable storeData, StoreData storeData) => storeData -> IO () registerInitialStore initial = do sdataE <- export initial storeE <- export . NewReactStoreHS =<< newMVar () js_createNewStore (storeJsKey (Proxy :: Proxy storeData)) sdataE storeE -- | First, 'transform' the store data according to the given action and then notify all registered -- controller-views to re-render themselves. -- -- Only a single thread can be transforming the store at any one time, so this function will block -- on an 'MVar' waiting for a previous transform to complete if one is in process. -- -- This function will 'error' if 'registerInitialStore' has not been called. transformStore :: forall storeData. StoreData storeData => Proxy storeData -> StoreAction storeData -> IO () transformStore _ action = do store :: NewReactStore storeData <- js_getNewStore (storeJsKey (Proxy :: Proxy storeData)) storeHS <- getNewStoreHS store modifyMVar_ (newStoreLock storeHS) $ \() -> do oldData :: storeData <- js_getNewStoreData store >>= unsafeDerefExport "transformStore" newData :: storeData <- transform action oldData js_updateNewStore store =<< export newData -- | Obtain the store data from a store. Note that the store data is stored in an MVar, so -- 'readStoreData' can block since it uses 'readMVar'. The 'MVar' is empty exactly when the store is -- being transformed, so there is a possiblity of deadlock if two stores try and access each other's -- data during transformation. -- -- This function will 'error' if 'registerInitialStore' has not been called. readStoreData :: forall storeData. (Typeable storeData, StoreData storeData) => IO storeData readStoreData = do store :: NewReactStore storeData <- js_getNewStore (storeJsKey (Proxy :: Proxy storeData)) js_getNewStoreData store >>= unsafeDerefExport "readStoreData" foreign import javascript unsafe "hsreact$storedata[$1]" js_getNewStore :: JSString -> IO (NewReactStore storeData) foreign import javascript unsafe "$1.hs" js_getNewStoreHS :: NewReactStore storeData -> IO (Export NewReactStoreHS) getNewStoreHS :: NewReactStore storeData -> IO NewReactStoreHS getNewStoreHS s = js_getNewStoreHS s >>= unsafeDerefExport "getNewStoreHS" {-# NOINLINE getNewStoreHS #-} foreign import javascript unsafe "$1.sdata" js_getNewStoreData :: NewReactStore storeData -> IO (Export storeData) foreign import javascript unsafe "hsreact$storedata[$1] = {sdata: $2, views: {}, hs: $3}" js_createNewStore :: JSString -> Export storeData -> Export NewReactStoreHS -> IO () -- | Perform the update, swapping the old export and the new export and then notifying the component views. foreign import javascript unsafe "hsreact$transform_new_store($1, $2)" js_updateNewStore :: NewReactStore storeData -> Export storeData -> IO ()
liqula/react-flux
src/React/Flux/Store.hs
bsd-3-clause
10,832
18
14
1,957
1,425
798
627
-1
-1
{-# LANGUAGE DeriveGeneric #-} {-# LANGUAGE NoImplicitPrelude #-} module HMenu.Command ( Command(..), execute, commandLine ) where import ClassyPrelude import Control.DeepSeq import Data.Binary import GHC.Generics (Generic) import System.Directory import System.Environment import System.FilePath import System.Posix.Process data Command = ShellCommand Text | GraphicalCommand Text deriving (Show, Eq, Ord, Generic) instance NFData Command instance Binary Command instance Hashable Command commandLine :: Command -> Text commandLine (ShellCommand c) = c commandLine (GraphicalCommand c) = c execute :: Command -> IO () execute c = do (cmd, args) <- toGraphical c executeFile (unpack cmd) True (map unpack args) Nothing toGraphical :: Command -> IO (Text, [Text]) toGraphical (ShellCommand c) = do t <- findTerminal case t of Nothing -> error "Cannot find a terminal emulator" Just t' -> return (t', ["-e", c]) toGraphical (GraphicalCommand c) = do s <- findShell return (s, ["-c", c]) findShell :: IO Text findShell = do s <- lookupEnv "SHELL" case s of Just s' -> return $ pack s' Nothing -> error "Empty $SHELL !" findTerminal :: IO (Maybe Text) findTerminal = do t <- lookupEnv "TERMINAL" firstExec $ case t of Just t' -> pack t' : defaultTerminals Nothing -> defaultTerminals where firstExec :: [Text] -> IO (Maybe Text) firstExec [] = return Nothing firstExec (f:fs) = do ex <- findExecutable $ unpack f case ex of Nothing -> firstExec fs Just f' -> return $ Just (pack f') defaultTerminals :: [Text] defaultTerminals = [ "x-terminal-emulator" , "urxvt" , "rxvt" , "termit" , "terminator" , "Eterm" , "aterm" , "xterm" , "gnome-terminal" , "roxterm" , "xfce4-terminal" , "termite" , "lxterminal" , "mate-terminal" , "terminology" , "st" , "qterminal" ]
Adirelle/hmenu
src/HMenu/Command.hs
bsd-3-clause
2,397
0
15
909
627
327
300
73
4
{-# LANGUAGE DeriveDataTypeable #-} module Main ( main ) where import System.IO import Data.List (sort) import qualified Data.ByteString.Lazy.Char8 as C import System.Console.CmdArgs import System.Directory (doesFileExist) data Options = Options { noSort :: Bool, file :: FilePath } deriving (Show,Data,Typeable) opts = Options { noSort = def &= help "Assume given line numbers is sorted.", file = def &= argPos 0 &= typFile } &= summary "getlines v0.1, (C) Naren Sundar 2010" &= program "getlines" &= details ["EXAMPLE: get lines 37 and 43 from file.", "$ echo \"37 43\" | getlines file","","", "http://github.com/nanonaren/getlines"] main = do options <- cmdArgs opts --check file exists exists <- doesFileExist (file options) if not exists then fail "Cannot open input file" else return () -- read lines lns <- C.lines `fmap` C.readFile (file options) lineNos <- fmap (map read.words) $ hGetContents stdin -- filter lines mapM_ C.putStrLn $ filterLines (if (noSort options) then lineNos else sort lineNos) lns filterLines ns ls = filterLines' ns (zip [1..] ls) filterLines' (n:ns) lns@((i,l):ls) | i > n = filterLines' ns lns | i == n = l : filterLines' ns ls | otherwise = filterLines' (n:ns) ls filterLines' [] _ = [] filterLines' _ [] = []
nanonaren/getlines
src/Main.hs
bsd-3-clause
1,416
0
12
369
449
235
214
36
3
module Text.HTML.Truncate(truncateHtml,truncateHtml',truncateStringLike) where import qualified Text.HTML.TagSoup as TS import qualified Text.StringLike as SL import Data.Char(isSpace) import Data.List(dropWhileEnd) {- Roughly, the algorithm works like this: 1. Parse the HTML to a list of tags 2. Walk through those tags. If a tag with actual text content is encountered, truncate it. If the text was not long enough to result in actually anything being truncated, keep going and truncate less the in the next tag that is encountered. Otherwise, just close all open tags and remove all remaining text in them. 3. Remove trailing empty tags -} -- | Truncate HTML, and ensure that tags are closed; Remove trailing empty tags truncateHtml :: SL.StringLike str => Int -> str -> str truncateHtml n txt = snd $ truncateHtml' n txt truncateHtml' :: SL.StringLike str => Int -> str -> (Int,str) truncateHtml' n txt = fmap (TS.renderTags . removeTrailingEmptyTags) $ go n 0 (TS.parseTags txt) where removeTrailingEmptyTags = removeTrailingEmptyTags' [] . reverse removeTrailingEmptyTags' accm (t@(TS.TagClose _) : ts) = removeTrailingEmptyTags' (t : accm) ts removeTrailingEmptyTags' accm (t@(TS.TagOpen _ _) : ts) = removeTrailingEmptyTags' (delCloseTag accm) ts removeTrailingEmptyTags' accm (t@(TS.TagText _) : ts) = reverse (t : ts) ++ accm removeTrailingEmptyTags' accm (t : ts) = removeTrailingEmptyTags' (t : accm) ts removeTrailingEmptyTags' accm [] = accm delCloseTag (t@(TS.TagClose _) : ts) = ts delCloseTag (t : ts) = t : delCloseTag ts delCloseTag [] = [] go :: (SL.StringLike str) => Int -- ^The number of remaining characters to truncate -> Int -- ^The number of open tags -> [TS.Tag str] -- ^The remaining tags to walk through -> (Int,[TS.Tag str]) go c openTags _ | c <= 0 && openTags <= 0 = (0,[]) -- we're done. Nothing to truncate, nothing to close go i _ [] = (i,[]) go 0 openTags (t@(TS.TagOpen _ _) : ts) = go 0 (openTags + 1) ts go c openTags (t@(TS.TagOpen _ _) : ts) = fmap (t :) (go c (openTags + 1) ts) go 0 openTags (t@(TS.TagClose _) : ts) = go 0 (openTags - 1) ts go c openTags (t@(TS.TagClose _) : ts) = fmap (t :) (go c (openTags - 1) ts) go 0 openTags ((TS.TagText str) : ts) = go 0 openTags ts go c openTags (t@(TS.TagText str) : ts) = case truncateStringLike c str of (c', str') -> fmap ((TS.TagText str') :) (go (max 0 c') openTags ts) go c openTags (t : ts) = fmap (t :) (go c openTags ts) -- | Truncate to full words. If actual truncation occured, remove the last (usually cut-off) word, then remove trailing whitespace. -- | Returns the truncated string and the number of characters that remain to be truncated truncateStringLike :: SL.StringLike str => Int -> str -> (Int, str) truncateStringLike c t = case truncateStringLike' c t of (0, t') -> (0, dropWhileEndSL isSpace $ dropWhileEndSL (not . isSpace) $ t') other -> other truncateStringLike' :: SL.StringLike str => Int -> str -> (Int, str) truncateStringLike' 0 t = (0, SL.empty) truncateStringLike' c t = case SL.uncons t of Nothing -> (c, t) Just (char, rest) -> fmap (SL.cons char) (truncateStringLike (c - 1) rest) -- note: could be optimized for double applications of this. dropWhileEndSL :: SL.StringLike a => (Char -> Bool) -> a -> a dropWhileEndSL p = SL.fromString . (dropWhileEnd p) . SL.toString
mruegenberg/html-truncate
Text/HTML/Truncate.hs
bsd-3-clause
3,618
0
14
877
1,153
612
541
44
15
{-# LANGUAGE OverloadedStrings #-} module Hrpg.Game.Resources.Zones.UpperCape ( upperCape ) where import Hrpg.Game.Resources.Mobs.SpawnTables (commonSpawns) import Hrpg.Framework.Zones.Zone upperCape = Zone (ZoneId 1000) "Upper Cape" "Woods on the bay" [] commonSpawns
cwmunn/hrpg
src/Hrpg/Game/Resources/Zones/UpperCape.hs
bsd-3-clause
285
0
7
44
59
37
22
6
1
-- This file contains utility functions not provided by the Haskell -- standard library. Most mirror functions in Harrison's lib.ml -- * Signature module Util.Lib ( time , timeIO , pow , funpow , decreasing -- Strings , substringIndex -- Map , (↦) , (⟾) ) where -- * Imports import Prelude import qualified Control.Exception as Exn import qualified Data.Map as Map import Data.Map (Map) import qualified System.CPUTime as Time import qualified Text.Printf as Printf -- * Misc timeIO :: IO a -> IO a timeIO a = do start <- Time.getCPUTime v <- a end <- Time.getCPUTime let diff :: Double = (fromIntegral (end - start)) / (1E12) _ <- Printf.printf "Computation time: %0.4f sec\n" diff return v time :: (a -> b) -> a -> IO b time f x = do start <- Time.getCPUTime y <- Exn.evaluate $ f x end <- Time.getCPUTime let diff :: Double = (fromIntegral (end - start)) / (1E12) _ <- Printf.printf "Computation time: %0.4f sec\n" diff return y pow' :: Int -> Int -> Int -> Int pow' _ 0 acc = acc pow' x n acc = pow' x (n-1) (x*acc) pow :: Int -> Int -> Int pow x y = if y >= 0 then pow' x y 1 else error "negative power" funpow :: Int -> (a -> a) -> a -> a funpow n f x = iterate f x !! n -- For use with sort decreasing :: Ord b => (a -> b) -> a -> a -> Ordering decreasing f x y = compare (f y) (f x) -- * Strings substringIndex :: String -> String -> Maybe Int substringIndex s1 s2 = let ind _ [] = Nothing ind n s = if take n s == s1 then Just n else ind (n+1) (tail s) in ind 0 s2 -- * Substitutions (↦) :: Ord a => a -> b -> Map a b -> Map a b (↦) = Map.insert (⟾) :: a -> b -> Map a b (⟾) = Map.singleton
etu-fkti5301-bgu/alt-exam_automated_theorem_proving
src/Util/Lib.hs
bsd-3-clause
1,698
0
14
428
697
367
330
-1
-1
module TypeDiff ( typeDiff -- exported for testing , sigMap , typeEq , alphaNormalize , normalizeConstrainNames ) where import Data.Char import Data.Maybe import Data.List import Data.Map as Map (Map) import qualified Data.Map as Map import Language.Haskell.Exts.Syntax import Language.Haskell.Exts.Parser import Data.Generics.Uniplate.Data typeDiff :: String -> String -> String typeDiff input1 input2 = unlines (missing ++ extra ++ wrongSigs) where sigs1 = sigMap input1 sigs2 = sigMap input2 names1 = Map.keys sigs1 names2 = Map.keys sigs2 missing = map format (names1 \\ names2) where format :: String -> String format = ("missing " ++) extra = map format (names2 \\ names1) where format :: String -> String format = ("extra " ++) wrongSigs :: [String] wrongSigs | null mismatches = [] | otherwise = "wrong types:" : mismatches where mismatches :: [String] mismatches = (catMaybes . map checkType . Map.toList) sigs1 checkType :: (String, String) -> Maybe String checkType (name, t1) = case Map.lookup name sigs2 of Just t2 | not (parseType_ t1 `typeEq` parseType_ t2) -> Just (format name t1 ++ "\n" ++ format name t2) _ -> Nothing format :: String -> String -> String format name type_ = " " ++ name ++ " :: " ++ type_ parseType_ :: String -> Type parseType_ type_ = case parseType type_ of ParseOk t -> t _ -> error ("can not parse type " ++ show type_) sigMap :: String -> Map String String sigMap = Map.fromList . map splitType . lines where splitType :: String -> (String, String) splitType = fmap stripSigMark . span (not . isSpace) stripSigMark :: String -> String stripSigMark = dropWhile isSpace . dropWhile (== ':') . dropWhile isSpace typeEq :: Type -> Type -> Bool typeEq t1 t2 = normalize t1 == normalize t2 where normalize = alphaNormalize . normalizeConstrainNames . normalizeConstrains . sortConstrains sortConstrains :: Type -> Type sortConstrains x = case x of TyForall a1 constrains a2 -> TyForall a1 (sort constrains) a2 _ -> x normalizeConstrains :: Type -> Type normalizeConstrains t = case t of TyForall a1 [ParenA a2] a3 -> TyForall a1 [a2] a3 _ -> t alphaNormalize :: Type -> Type alphaNormalize t = transformBi f t where f :: Name -> Name f name = fromMaybe name $ lookup name mapping names :: [Name] names = (nub . filter isTyVar . universeBi) t isTyVar :: Name -> Bool isTyVar x = case x of Ident n -> null (takeWhile isUpper n) _ -> False mapping :: [(Name, Name)] mapping = zip names vars vars :: [Name] vars = map (Ident . ('t' :) . show) [0 :: Integer ..] normalizeConstrainNames :: Type -> Type normalizeConstrainNames t = transformBi f t where f :: QName -> QName f name = case name of Qual (ModuleName "GHC.Base") (Ident "Applicative") -> Qual (ModuleName "Control.Applicative") (Ident "Applicative") Qual (ModuleName "GHC.Base") (Ident "Alternative") -> Qual (ModuleName "Control.Applicative") (Ident "Alternative") Qual (ModuleName "GHC.Base") (Ident "MonadPlus") -> Qual (ModuleName "Control.Monad") (Ident "MonadPlus") Qual (ModuleName "GHC.Base") (Ident "Maybe") -> Qual (ModuleName "Data.Maybe") (Ident "Maybe") Qual (ModuleName "GHC.Base") (Ident "Monoid") -> Qual (ModuleName "Data.Monoid") (Ident "Monoid") Qual (ModuleName "GHC.Types") (Ident "Bool") -> Qual (ModuleName "GHC.Bool") (Ident "Bool") Qual (ModuleName "GHC.Types") (Ident "Ordering") -> Qual (ModuleName "GHC.Ordering") (Ident "Ordering") _ -> name
beni55/base-compat
typediff/src/TypeDiff.hs
mit
3,765
0
18
937
1,283
661
622
85
8
{- | Module : ./atermlib/src/ATerm/Diff.hs Description : compute the differences of two unshared ATerms Copyright : (c) Klaus Luettich, Uni Bremen 2005 License : GPLv2 or higher, see LICENSE.txt Maintainer : Christian.Maeder@dfki.de Stability : provisional Portability : non-portable (imports ATerm.Unshared) Provides functions that calculate differences in unshared 'ATerm's. -} module ATerm.Diff (atDiff) where import ATerm.Unshared import Data.List {- | all diferences between both terms are replaced by appropiate placeholders (in @\<\>@) and the differing terms are added to the list of ATerm as arguments to the function symbol @diff@. -} -- {- /Note: this function ignores annotions and the resulting ATerm does not contain any annotation!/ -} atDiff :: ATerm -> ATerm -> (ATerm, [ATerm]) atDiff a1@(AAppl s1 atl1 _) a2@(AAppl s2 atl2 _) | s1 == s2 && atl1 == atl2 = (AAppl s1 atl1 [], []) | s1 == s2 && length atl1 == length atl2 = case atDiffL atl1 atl2 of (diffs, atl) -> (AAppl s1 atl [], diffs) | otherwise = (AAppl "<diff-appls>" [] [], [AAppl "diff" [a1, a2] []]) atDiff a1@(AInt i1 _) a2@(AInt i2 _) | i1 == i2 = (AInt i1 [], []) | otherwise = (AAppl "<diff-int>" [] [], [AAppl "diff" [a1, a2] []]) atDiff a1@(AList l1 _) a2@(AList l2 _) | l1 == l2 = (AList l1 [], []) | length l1 == length l2 = case atDiffL l1 l2 of (diffs, atl) -> (AList atl [], diffs) | otherwise = (AList [AAppl "<diff-lists>" [] []] [], [AAppl "diff" [a1, a2] []]) atDiff a1 a2 = (AAppl "<diff-types>" [] [], [AAppl "diff" [a1, a2] []]) atDiffL :: [ATerm] -> [ATerm] -> ([ATerm], [ATerm]) atDiffL atl1 atl2 = mapAccumL (\ acc (ia1, ia2) -> case atDiff ia1 ia2 of (at, diffs) -> (acc ++ diffs, at)) [] (zip atl1 atl2)
spechub/Hets
atermlib/src/ATerm/Diff.hs
gpl-2.0
1,866
0
12
459
640
338
302
27
1
{-# OPTIONS -O2 -Wall #-} {-# LANGUAGE GeneralizedNewtypeDeriving #-} module Data.Store.IRef (IRef, guid, unsafeFromGuid, anchor) where import Data.Binary (Binary) import Data.Store.Guid (Guid) import qualified Data.Store.Guid as Guid newtype IRef a = IRef { guid :: Guid } deriving (Eq, Ord, Binary, Read, Show) -- Wrapper modules need to create an IRef unsafeFromGuid :: Guid -> IRef a unsafeFromGuid = IRef anchor :: (Binary a) => String -> IRef a anchor = unsafeFromGuid . Guid.fromString
alonho/bottle
src/Data/Store/IRef.hs
gpl-3.0
532
0
7
113
140
85
55
14
1
{-| Module : Util.DynamicLinker Description : Platform-specific dynamic linking support. Add new platforms to this file through conditional compilation. Copyright : License : BSD3 Maintainer : The Idris Community. -} {-# LANGUAGE ExistentialQuantification, CPP, ScopedTypeVariables #-} module Util.DynamicLinker ( ForeignFun(..) , DynamicLib(..) , tryLoadLib , tryLoadFn ) where #ifdef IDRIS_FFI import Foreign.LibFFI import Foreign.Ptr (Ptr(), nullPtr, FunPtr, nullFunPtr, castPtrToFunPtr) import System.Directory #ifndef mingw32_HOST_OS import Control.Exception (try, IOException, throwIO) import Data.Array (Array, inRange, bounds, (!)) import Data.Functor ((<$>)) import Data.Maybe (catMaybes) import System.Posix.DynamicLinker import System.FilePath.Posix ((</>)) import Text.Regex.TDFA #else import qualified Control.Exception as Exception (catch, IOException) import System.Win32.DLL import System.Win32.Types import System.FilePath.Windows ((</>)) type DL = HMODULE #endif hostDynamicLibExt :: String #if defined(linux_HOST_OS) || defined(freebsd_HOST_OS) \ || defined(dragonfly_HOST_OS) || defined(openbsd_HOST_OS) \ || defined(netbsd_HOST_OS) hostDynamicLibExt = "so" #elif defined(darwin_HOST_OS) hostDynamicLibExt = "dylib" #elif defined(mingw32_HOST_OS) hostDynamicLibExt = "dll" #else hostDynamicLibExt = error $ unwords [ "Undefined file extension for dynamic libraries" , "in Idris' Util.DynamicLinker." ] #endif data ForeignFun = forall a. Fun { fun_name :: String , fun_handle :: FunPtr a } data DynamicLib = Lib { lib_name :: String , lib_handle :: DL } instance Eq DynamicLib where (Lib a _) == (Lib b _) = a == b firstExisting :: [FilePath] -> IO (Maybe FilePath) firstExisting [] = return Nothing firstExisting (f:fs) = do exists <- doesFileExist f if exists then return (Just f) else firstExisting fs libFileName :: [FilePath] -> String -> IO String libFileName dirs lib = do let names = [lib, lib ++ "." ++ hostDynamicLibExt] cwd <- getCurrentDirectory found <- firstExisting $ map ("."</>) names ++ [d </> f | d <- cwd:dirs, f <- names] return $ maybe (lib ++ "." ++ hostDynamicLibExt) id found #ifndef mingw32_HOST_OS -- Load a dynamic library on POSIX systems. -- In the simple case, we just find the appropriate filename and call dlopen(). -- In the complicated case our "foo.so" isn't actually a library. Some of the -- .so files on modern Linux systems are linker scripts instead. dlopen() -- doesn't know anything about those. We need to look inside the script for the -- actual library path and load that. This is a horrible hack, the correct -- method would be to actually parse the scripts and execute them. The approach -- below is what GHC does. tryLoadLib :: [FilePath] -> String -> IO (Maybe DynamicLib) tryLoadLib dirs lib = do filename <- libFileName dirs lib res :: Either IOException DL <- try $ dlopen filename [RTLD_NOW, RTLD_GLOBAL] mbDL <- case res of Right handle -> return $ Just handle #ifdef linux_HOST_OS Left ex -> -- dlopen failed, run a regex to see if the error message looks like it -- could be a linker script. case matchAllText invalidLibRegex (show ex) of (x:_) -> do if inRange (bounds x) 1 then do -- filename above may be a relative path. Get the full path out of -- the error message. let realPath = fst $ x ! 1 fileLines <- lines <$> readFile realPath -- Go down the linker script line by line looking for .so -- filenames and try each one. let matches = catMaybes $ map (getLastMatch . matchAllText linkerScriptRegex) fileLines mapMFirst (\f -> dlopen f [RTLD_NOW, RTLD_GLOBAL]) matches else return Nothing [] -> return Nothing #else Left ex -> throwIO ex #endif case mbDL of Just handle -> if undl handle == nullPtr then return Nothing else return . Just $ Lib lib handle Nothing -> return Nothing getLastMatch :: [MatchText String] -> Maybe String getLastMatch [] = Nothing getLastMatch (x:_) = case bounds x of (low, high) -> if low > high then Nothing else Just $ fst $ x ! high mapMFirst :: (a -> IO b) -> [a] -> IO (Maybe b) mapMFirst f [] = return Nothing mapMFirst f (a:as) = do res <- try (f a) case res of Left (ex :: IOException) -> mapMFirst f as Right res -> return $ Just res -- Both regexes copyright 2009-2011 Howard B. Golden, CJ van den Berg and Ian -- Lynagh. From the Glasgow Haskell Compiler. BSD licensed. invalidLibRegex :: Regex invalidLibRegex = makeRegex "(([^ \t()])+\\.so([^ \t:()])*):([ \t])*(invalid ELF header|file too short)" linkerScriptRegex :: Regex linkerScriptRegex = makeRegex "(GROUP|INPUT) *\\( *([^ )]+)" tryLoadFn :: String -> DynamicLib -> IO (Maybe ForeignFun) tryLoadFn fn (Lib _ h) = do cFn <- dlsym h fn if cFn == nullFunPtr then return Nothing else return . Just $ Fun fn cFn #else tryLoadLib :: [FilePath] -> String -> IO (Maybe DynamicLib) tryLoadLib dirs lib = do filename <- libFileName dirs lib handle <- Exception.catch (loadLibrary filename) nullPtrOnException if handle == nullPtr then return Nothing else return . Just $ Lib lib handle where nullPtrOnException :: Exception.IOException -> IO DL nullPtrOnException e = return nullPtr -- `show e` will however give broken error message tryLoadFn :: String -> DynamicLib -> IO (Maybe ForeignFun) tryLoadFn fn (Lib _ h) = do cFn <- getProcAddress h fn if cFn == nullPtr then return Nothing else return . Just $ Fun fn (castPtrToFunPtr cFn) #endif #else -- no libffi, just add stubbs. data DynamicLib = Lib { lib_name :: String , lib_handle :: () } deriving Eq data ForeignFun = forall a. Fun { fun_name :: String , fun_handle :: () } tryLoadLib :: [FilePath] -> String -> IO (Maybe DynamicLib) tryLoadLib fps lib = do putStrLn $ "WARNING: Cannot load '" ++ lib ++ "' at compile time because Idris was compiled without libffi support." return Nothing tryLoadFn :: String -> DynamicLib -> IO (Maybe ForeignFun) tryLoadFn fn lib = do putStrLn $ "WARNING: Cannot load '" ++ fn ++ "' at compile time because Idris was compiled without libffi support." return Nothing #endif
ozgurakgun/Idris-dev
src/Util/DynamicLinker.hs
bsd-3-clause
7,268
0
26
2,278
1,138
604
534
16
1
-- Copyright (c) 2015 Eric McCorkle. All rights reserved. -- -- Redistribution and use in source and binary forms, with or without -- modification, are permitted provided that the following conditions -- are met: -- -- 1. Redistributions of source code must retain the above copyright -- notice, this list of conditions and the following disclaimer. -- -- 2. Redistributions in binary form must reproduce the above copyright -- notice, this list of conditions and the following disclaimer in the -- documentation and/or other materials provided with the distribution. -- -- 3. Neither the name of the author nor the names of any contributors -- may be used to endorse or promote products derived from this software -- without specific prior written permission. -- -- THIS SOFTWARE IS PROVIDED BY THE AUTHORS AND CONTRIBUTORS ``AS IS'' -- AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED -- TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A -- PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHORS -- OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -- SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -- LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF -- USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND -- ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, -- OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT -- OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF -- SUCH DAMAGE. module Tests.Language.Salt.Surface(tests) where import Test.HUnitPlus.Base tests :: Test tests = "Surface" ~: ([] :: [Test])
emc2/saltlang
test/library/Tests/Language/Salt/Surface.hs
bsd-3-clause
1,684
0
7
285
73
57
16
4
1
{- | Module : $Header$ Description : Instance of class Logic for propositional logic Copyright : (c) Dominik Luecke, Uni Bremen 2007 License : GPLv2 or higher, see LICENSE.txt Maintainer : luecke@informatik.uni-bremen.de Stability : experimental Portability : non-portable (imports Logic.Logic) Test cases for Prop2CNF -} {- Ref. http://en.wikipedia.org/wiki/Propositional_logic Till Mossakowski, Joseph Goguen, Razvan Diaconescu, Andrzej Tarlecki. What is a Logic?. In Jean-Yves Beziau (Ed.), Logica Universalis, pp. 113-@133. Birkhaeuser. 2005. -} module Propositional.TestProp2CNF where import Propositional.Prop2CNF -- FIXME: What happened to Prop2CNF? import Propositional.AS_BASIC_Propositional import Propositional.Sign import Common.Id import Common.AS_Annotation import qualified SoftFOL.ProverState as PState import qualified Propositional.Conversions as PC import Propositional.Prove import Propositional.ProverState import qualified SoftFOL.Sign as SPS aId :: Id aId = stringToId "a" bId :: Id bId = stringToId "b" cId :: Id cId = stringToId "c" mySig :: Sign mySig = addToSig (addToSig (addToSig emptySig aId) bId) cId myForm :: FORMULA myForm = Disjunction [Predication (mkSimpleId "a"), Negation (Predication (mkSimpleId "a")) nullRange] nullRange myOtherForm :: FORMULA myOtherForm = Conjunction [Equivalence (Predication (mkSimpleId "a")) (Predication (mkSimpleId "c")) nullRange] nullRange {- myForm :: FORMULA myForm = (Predication (mkSimpleId "a")) -} myEmptyForm = makeNamed "myForm" myForm otherForm = makeNamed "myOtherForm" myOtherForm myForms = [myEmptyForm, otherForm] runAll = translateToCNF (mySig, [myEmptyForm]) showStuff :: IO String showStuff = PC.showDIMACSProblem "Problem " mySig myForms [] showProof = PC.goalDIMACSProblem "DIMACSProblem" (propProverState mySig myForms []) otherForm []
keithodulaigh/Hets
Propositional/TestProp2CNF.hs
gpl-2.0
1,974
0
11
374
339
188
151
36
1
{-# language ScopedTypeVariables, DeriveDataTypeable, NamedFieldPuns, MultiParamTypeClasses, FlexibleInstances #-} -- | objects that seem like tiles, but move when Nikki touches them module Sorts.FallingTiles (sorts) where import Data.Data import Data.Abelian import Data.Set (member) import Data.IORef import System.FilePath import System.Random import qualified Physics.Hipmunk as H import Physics.Chipmunk as CM hiding (Static) import Graphics.Qt import Utils import Base -- * configuration -- | Tiles that are available as falling tiles. Same format as in Sorts/Tiles.hs names = [ ("tiles/black-falling", Position 1 1, Size 64 64), ("tiles/white-falling", Position 1 1, Size 64 64), ("tiles/yellow-falling", Position 1 1, Size 64 64), ("tiles/green-falling", Position 1 1, Size 64 64), ("tiles/aqua-falling", Position 1 1, Size 64 64), ("tiles/blue-falling", Position 1 1, Size 64 64), ("tiles/pink-falling", Position 1 1, Size 64 64), ("tiles/red-falling", Position 1 1, Size 64 64) ] -- | time in seconds, before the tiles start to fall after touching Nikki timeBeforeGettingLoose :: Seconds timeBeforeGettingLoose = 0.4 fallingTilesMaterialMass = 0.5 -- * loading sorts :: [RM (Maybe Sort_)] sorts = map ((Just <$>) . (\ (a, b, c) -> mkSort a b c)) names mkSort :: String -> Offset Int -> Size Double -> RM Sort_ mkSort name offset size = do pngFile <- getDataFileName (pngDir </> name <.> "png") Sort_ <$> (TSort name <$> loadPixmap (fmap fromIntegral offset) size pngFile <*> io (newIORef (mkStdGen 42))) data TSort = TSort { name :: String, tilePixmap :: Pixmap, randomGenRef :: IORef StdGen } deriving (Show, Typeable) instance Show (IORef StdGen) where show = const "<IORef StdGen>" randomAngVel :: TSort -> IO Angle randomAngVel sort = do randomGen <- readIORef $ randomGenRef sort let (angVel, randomGen') = randomR (-2, 2) randomGen writeIORef (randomGenRef sort) randomGen' return angVel data FallingTile = FallingTile { tileAttributes :: BodyAttributes, chipmunk :: Chipmunk, status :: Status } deriving (Show, Typeable) data Status = Static | GettingLoose Seconds | Loose deriving (Show, Typeable) instance Sort TSort FallingTile where sortId TSort{name} = SortId ("fallingTile/" ++ name) size (TSort _ pix _) = pixmapSize pix renderIconified sort ptr = do renderPixmapSimple ptr (tilePixmap sort) let Size w h = fmap (subtract 1) $ size sort -- draw a red cross on top setPenColor ptr red 2 drawLine ptr zero (Position w h) drawLine ptr (Position w 0) (Position 0 h) initialize app _ (Just space) sort@TSort{} editorPosition Nothing _ = io $ do (chip, attributes) <- initializeBox space sort editorPosition modifyApplyForce chip (CM.scale (Vector 0 (- gravity)) staticMass) return $ FallingTile attributes chip Static initialize app _ Nothing sort editorPosition Nothing _ = io $ do let (_, baryCenterOffset) = mkShape sort position = epToPosition (size sort) editorPosition vector = position2vector position +~ baryCenterOffset chip = ImmutableChipmunk position 0 baryCenterOffset [] return $ FallingTile (StaticBodyAttributes vector) chip Static immutableCopy t@FallingTile{chipmunk} = CM.immutableCopy chipmunk >>= \ x -> return t{chipmunk = x} chipmunks (FallingTile _ c _) = [c] isUpdating = const True updateNoSceneChange sort _ config _ mode now contacts cd fallingTile = case status fallingTile of Static -> if any (`member` fallingTiles contacts) (shapes (chipmunk fallingTile)) then return fallingTile{status = GettingLoose now} else return fallingTile GettingLoose t -> if now - t >= timeBeforeGettingLoose then do modifyApplyOnlyForce (chipmunk fallingTile) zero let b = body $ chipmunk fallingTile H.mass b $= (mass $ tileAttributes fallingTile) moment b $= (inertia_ $ tileAttributes fallingTile) angVel <- randomAngVel sort modifyAngVel (chipmunk fallingTile) (const angVel) return $ fallingTile{status = Loose} else return fallingTile Loose -> return fallingTile renderObject _ _ t@FallingTile{chipmunk = ImmutableChipmunk{}} sort@TSort{tilePixmap} ptr offset _now = do (position, rad) <- getRenderPositionAndAngle $ chipmunk t return [RenderPixmap tilePixmap position (Just rad)] initializeBox :: Space -> TSort -> EditorPosition -> IO (Chipmunk, BodyAttributes) initializeBox space sort ep = do let (shape, baryCenterOffset) = mkShape sort shapeWithAttributes = (mkShapeDescription shapeAttributes shape) pos :: Vector pos = position2vector (epToPosition (size sort) ep) +~ baryCenterOffset bodyAttributes = mkMaterialBodyAttributes fallingTilesMaterialMass [shape] pos -- this is a hack: mass and inertia set to very large values to simulate static tiles staticBodyAttributes = bodyAttributes{mass = staticMass, inertia_ = staticInertia} chip <- initChipmunk space staticBodyAttributes [shapeWithAttributes] baryCenterOffset return (chip, bodyAttributes) staticMass = 1000000000000 staticInertia = 1000000000000000000 mkShape :: TSort -> (ShapeType, Vector) mkShape sort = (box, baryCenterOffset) where box = Polygon [ upperLeft +~ xUnit, upperLeft +~ yUnit, lowerLeft -~ yUnit, lowerLeft +~ xUnit, lowerRight -~ xUnit, lowerRight -~ yUnit, upperRight +~ yUnit, upperRight -~ xUnit] xUnit = Vector 1 0 yUnit = Vector 0 1 Size w h :: Size CpFloat = fmap realToFrac $ size sort wh = w / 2 hh = h / 2 baryCenterOffset = Vector wh hh -- | falling tiles have to be smaller than normal -- to avoid them to be wedged in between normal tiles. -- Except for the upper edge, to allow smooth walking... low = hh - 1 up = (- hh) -- + 1 left = (- wh) + 1 right = wh - 1 upperLeft = Vector left up lowerLeft = Vector left low lowerRight = Vector right low upperRight = Vector right up shapeAttributes :: ShapeAttributes shapeAttributes = ShapeAttributes { elasticity = 0.5, friction = 2.0, collisionType = FallingTileCT }
changlinli/nikki
src/Sorts/FallingTiles.hs
lgpl-3.0
6,716
0
16
1,818
1,900
990
910
144
1
module Turbinado.Environment.Files( getFile, getFile_u, getFileContent, getFileDecode ) where import qualified Data.ByteString as BS import qualified Data.Map as M import Data.Maybe import Network.HTTP import Network.HTTP.Headers import Network.URI import Codec.MIME.Type import Codec.MIME.Decode import Turbinado.Environment.Params import Turbinado.Environment.Types import Turbinado.Utility.Data -- | Attempt to get a File from the POST getFile :: (HasEnvironment m) => String -> m (Maybe MIMEValue) getFile f = do populateParamsAndFiles e <- getEnvironment let Files fs = fromJust' "Turbinado.Environment.Files.getFile: Files is Nothing" $ getFiles e return $ M.lookup f fs -- | An unsafe version of getFile. Errors if the key does not exist. getFile_u :: (HasEnvironment m) => String -> m MIMEValue getFile_u f = do r <- getFile f maybe (error $ "getFile_u : key does not exist - \"" ++ f ++ "\"") return r getFileContent :: MIMEValue -> String getFileContent mv = case (mime_val_content mv) of Single c -> c _ -> error "Turbinado.Environment.Params.getContent: called with a Multi Content" getFileDecode :: MIMEValue -> String getFileDecode mv = case (mime_val_content mv) of --Single c -> decodeBody "base64" c Single c -> decodeWords c _ -> error "Turbinado.Environment.Params.getContent: called with a Multi Content"
alsonkemp/turbinado-website
Turbinado/Environment/Files.hs
bsd-3-clause
1,530
0
11
391
330
175
155
35
2
{-# LANGUAGE NoImplicitPrelude #-} {-# LANGUAGE TypeSynonymInstances #-} -- | The internal FFI module. module Fay.FFI (Fay ,Nullable (..) ,Defined (..) ,Ptr ,Automatic ,ffi) where import Data.String (IsString) import Fay.Types import Prelude (error) -- | Values that may be null -- Nullable x decodes to x, Null decodes to null. data Nullable a = Nullable a | Null -- | Values that may be undefined -- Defined x encodes to x, Undefined decodes to undefined. -- An undefined property in a record will be removed when encoding. data Defined a = Defined a | Undefined -- | Do not serialize the specified type. This is useful for, e.g. -- -- > foo :: String -> String -- > foo = ffi "%1" -- -- This would normally serialize and unserialize the string, for no -- reason, in this case. Instead: -- -- > foo :: Ptr String -> Ptr String -- -- Will just give an identity function. type Ptr a = a -- | The opposite of "Ptr". Serialize the specified polymorphic type. -- -- > foo :: Automatic a -> String -- type Automatic a = a -- | Declare a foreign action. ffi :: IsString s => s -- ^ The foreign value. -> a -- ^ Bottom. ffi = error "Fay.FFI.ffi: Used foreign function outside a JS engine context."
beni55/fay
src/Fay/FFI.hs
bsd-3-clause
1,284
0
6
315
147
101
46
20
1
{-# LANGUAGE FlexibleContexts #-} {-# LANGUAGE MultiParamTypeClasses #-} {-# LANGUAGE OverloadedStrings #-} {-# LANGUAGE TemplateHaskell #-} module Stack.Upgrade (upgrade) where import Control.Monad.Catch import Control.Monad.IO.Class import Control.Monad.Logger import Control.Monad.Reader (MonadReader, asks) import Control.Monad.Trans.Control import Data.Foldable (forM_) import qualified Data.Map as Map import qualified Data.Set as Set import Network.HTTP.Client.Conduit (HasHttpManager, getHttpManager) import Path import qualified Paths_stack as Paths import Stack.Build import Stack.Build.Types import Stack.Config import Stack.Fetch import Stack.PackageIndex import Stack.Setup import Stack.Types import Stack.Types.Internal import Stack.Types.StackT import System.IO.Temp (withSystemTempDirectory) import System.Process.Run upgrade :: (MonadIO m, MonadMask m, MonadReader env m, HasConfig env, HasHttpManager env, MonadLogger m, HasTerminal env, HasLogLevel env, MonadBaseControl IO m) => Bool -- ^ use Git? -> Maybe Resolver -> m () upgrade fromGit mresolver = withSystemTempDirectory "stack-upgrade" $ \tmp' -> do menv <- getMinimalEnvOverride tmp <- parseAbsDir tmp' mdir <- if fromGit then do $logInfo "Cloning stack" runIn tmp "git" menv [ "clone" , "https://github.com/commercialhaskell/stack" -- TODO allow to be configured , "stack" , "--depth" , "1" ] Nothing return $ Just $ tmp </> $(mkRelDir "stack") else do updateAllIndices menv caches <- getPackageCaches menv let latest = Map.fromListWith max $ map toTuple $ Map.keys -- Mistaken upload to Hackage, just ignore it $ Map.delete (PackageIdentifier $(mkPackageName "stack") $(mkVersion "9.9.9")) caches case Map.lookup $(mkPackageName "stack") latest of Nothing -> error "No stack found in package indices" Just version | version <= fromCabalVersion Paths.version -> do $logInfo "Already at latest version, no upgrade required" return Nothing Just version -> do let ident = PackageIdentifier $(mkPackageName "stack") version paths <- unpackPackageIdents menv tmp Nothing $ Set.singleton ident case Map.lookup ident paths of Nothing -> error "Stack.Upgrade.upgrade: invariant violated, unpacked directory not found" Just path -> return $ Just path manager <- asks getHttpManager logLevel <- asks getLogLevel terminal <- asks getTerminal configMonoid <- asks $ configConfigMonoid . getConfig forM_ mdir $ \dir -> liftIO $ do bconfig <- runStackLoggingT manager logLevel terminal $ do lc <- loadConfig configMonoid (Just $ dir </> $(mkRelFile "stack.yaml")) lcLoadBuildConfig lc mresolver ThrowException envConfig1 <- runStackT manager logLevel bconfig terminal setupEnv runStackT manager logLevel envConfig1 terminal $ build (const $ return ()) defaultBuildOpts { boptsTargets = ["stack"] , boptsInstallExes = True }
GaloisInc/stack
src/Stack/Upgrade.hs
bsd-3-clause
3,964
0
24
1,533
792
400
392
81
5
{-# LANGUAGE Trustworthy #-} {-# LANGUAGE CPP, MagicHash #-} {-# OPTIONS_HADDOCK hide #-} module GHC.List ( mylen ) where import GHC.Base hiding (assert) {-# INLINE mfoldr #-} mfoldr :: (a -> b -> b) -> b -> [a] -> b mfoldr k z = go where go [] = z go (y:ys) = y `k` go ys {-@ assert mylen :: xs: [a] -> {v: Int | v = len(xs)} @-} mylen :: [a] -> Int mylen = mfoldr (\_ -> (1 +)) 0 --{-# INLINE mmap #-} --mmap f = go -- where go [] = [] -- go (x:xs) = (f x) : (go xs) -- --myadd :: [Int] -> [Int] --myadd = mmap (1 +)
mightymoose/liquidhaskell
benchmarks/ghc-7.4.1/List-mini.hs
bsd-3-clause
567
0
9
166
145
87
58
13
2
<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE helpset PUBLIC "-//Sun Microsystems Inc.//DTD JavaHelp HelpSet Version 2.0//EN" "http://java.sun.com/products/javahelp/helpset_2_0.dtd"> <helpset version="2.0" xml:lang="ko-KR"> <title>SOAP Support Add-on</title> <maps> <homeID>soap</homeID> <mapref location="map.jhm"/> </maps> <view> <name>TOC</name> <label>Contents</label> <type>org.zaproxy.zap.extension.help.ZapTocView</type> <data>toc.xml</data> </view> <view> <name>Index</name> <label>Index</label> <type>javax.help.IndexView</type> <data>index.xml</data> </view> <view> <name>Search</name> <label>Search</label> <type>javax.help.SearchView</type> <data engine="com.sun.java.help.search.DefaultSearchEngine"> JavaHelpSearch </data> </view> <view> <name>Favorites</name> <label>Favorites</label> <type>javax.help.FavoritesView</type> </view> </helpset>
kingthorin/zap-extensions
addOns/soap/src/main/javahelp/org/zaproxy/zap/extension/soap/resources/help_ko_KR/helpset_ko_KR.hs
apache-2.0
965
77
67
157
413
209
204
-1
-1
module Test1 where f = e2 where e2 = 1 - 2
kmate/HaRe
old/testing/refacSlicing/Test1AST.hs
bsd-3-clause
44
0
7
13
20
12
8
2
1
{- (c) The AQUA Project, Glasgow University, 1993-1998 \section[Simplify]{The main module of the simplifier} -} {-# LANGUAGE CPP #-} module Simplify ( simplTopBinds, simplExpr, simplRules ) where #include "HsVersions.h" import DynFlags import SimplMonad import Type hiding ( substTy, extendTvSubst, substTyVar ) import SimplEnv import SimplUtils import FamInstEnv ( FamInstEnv ) import Literal ( litIsLifted ) --, mkMachInt ) -- temporalily commented out. See #8326 import Id import MkId ( seqId, voidPrimId ) import MkCore ( mkImpossibleExpr, castBottomExpr ) import IdInfo import Name ( Name, mkSystemVarName, isExternalName ) import Coercion hiding ( substCo, substTy, substCoVar, extendTvSubst ) import OptCoercion ( optCoercion ) import FamInstEnv ( topNormaliseType_maybe ) import DataCon ( DataCon, dataConWorkId, dataConRepStrictness , isMarkedStrict ) --, dataConTyCon, dataConTag, fIRST_TAG ) --import TyCon ( isEnumerationTyCon ) -- temporalily commented out. See #8326 import CoreMonad ( Tick(..), SimplifierMode(..) ) import CoreSyn import Demand ( StrictSig(..), dmdTypeDepth, isStrictDmd ) import PprCore ( pprCoreExpr ) import CoreUnfold import CoreUtils import CoreArity --import PrimOp ( tagToEnumKey ) -- temporalily commented out. See #8326 import Rules ( mkSpecInfo, lookupRule, getRules ) import TysPrim ( voidPrimTy ) --, intPrimTy ) -- temporalily commented out. See #8326 import BasicTypes ( TopLevelFlag(..), isTopLevel, RecFlag(..) ) import MonadUtils ( foldlM, mapAccumLM, liftIO ) import Maybes ( orElse ) --import Unique ( hasKey ) -- temporalily commented out. See #8326 import Control.Monad import Outputable import FastString import Pair import Util import ErrUtils {- The guts of the simplifier is in this module, but the driver loop for the simplifier is in SimplCore.hs. ----------------------------------------- *** IMPORTANT NOTE *** ----------------------------------------- The simplifier used to guarantee that the output had no shadowing, but it does not do so any more. (Actually, it never did!) The reason is documented with simplifyArgs. ----------------------------------------- *** IMPORTANT NOTE *** ----------------------------------------- Many parts of the simplifier return a bunch of "floats" as well as an expression. This is wrapped as a datatype SimplUtils.FloatsWith. All "floats" are let-binds, not case-binds, but some non-rec lets may be unlifted (with RHS ok-for-speculation). ----------------------------------------- ORGANISATION OF FUNCTIONS ----------------------------------------- simplTopBinds - simplify all top-level binders - for NonRec, call simplRecOrTopPair - for Rec, call simplRecBind ------------------------------ simplExpr (applied lambda) ==> simplNonRecBind simplExpr (Let (NonRec ...) ..) ==> simplNonRecBind simplExpr (Let (Rec ...) ..) ==> simplify binders; simplRecBind ------------------------------ simplRecBind [binders already simplfied] - use simplRecOrTopPair on each pair in turn simplRecOrTopPair [binder already simplified] Used for: recursive bindings (top level and nested) top-level non-recursive bindings Returns: - check for PreInlineUnconditionally - simplLazyBind simplNonRecBind Used for: non-top-level non-recursive bindings beta reductions (which amount to the same thing) Because it can deal with strict arts, it takes a "thing-inside" and returns an expression - check for PreInlineUnconditionally - simplify binder, including its IdInfo - if strict binding simplStrictArg mkAtomicArgs completeNonRecX else simplLazyBind addFloats simplNonRecX: [given a *simplified* RHS, but an *unsimplified* binder] Used for: binding case-binder and constr args in a known-constructor case - check for PreInLineUnconditionally - simplify binder - completeNonRecX ------------------------------ simplLazyBind: [binder already simplified, RHS not] Used for: recursive bindings (top level and nested) top-level non-recursive bindings non-top-level, but *lazy* non-recursive bindings [must not be strict or unboxed] Returns floats + an augmented environment, not an expression - substituteIdInfo and add result to in-scope [so that rules are available in rec rhs] - simplify rhs - mkAtomicArgs - float if exposes constructor or PAP - completeBind completeNonRecX: [binder and rhs both simplified] - if the the thing needs case binding (unlifted and not ok-for-spec) build a Case else completeBind addFloats completeBind: [given a simplified RHS] [used for both rec and non-rec bindings, top level and not] - try PostInlineUnconditionally - add unfolding [this is the only place we add an unfolding] - add arity Right hand sides and arguments ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In many ways we want to treat (a) the right hand side of a let(rec), and (b) a function argument in the same way. But not always! In particular, we would like to leave these arguments exactly as they are, so they will match a RULE more easily. f (g x, h x) g (+ x) It's harder to make the rule match if we ANF-ise the constructor, or eta-expand the PAP: f (let { a = g x; b = h x } in (a,b)) g (\y. + x y) On the other hand if we see the let-defns p = (g x, h x) q = + x then we *do* want to ANF-ise and eta-expand, so that p and q can be safely inlined. Even floating lets out is a bit dubious. For let RHS's we float lets out if that exposes a value, so that the value can be inlined more vigorously. For example r = let x = e in (x,x) Here, if we float the let out we'll expose a nice constructor. We did experiments that showed this to be a generally good thing. But it was a bad thing to float lets out unconditionally, because that meant they got allocated more often. For function arguments, there's less reason to expose a constructor (it won't get inlined). Just possibly it might make a rule match, but I'm pretty skeptical. So for the moment we don't float lets out of function arguments either. Eta expansion ~~~~~~~~~~~~~~ For eta expansion, we want to catch things like case e of (a,b) -> \x -> case a of (p,q) -> \y -> r If the \x was on the RHS of a let, we'd eta expand to bring the two lambdas together. And in general that's a good thing to do. Perhaps we should eta expand wherever we find a (value) lambda? Then the eta expansion at a let RHS can concentrate solely on the PAP case. ************************************************************************ * * \subsection{Bindings} * * ************************************************************************ -} simplTopBinds :: SimplEnv -> [InBind] -> SimplM SimplEnv simplTopBinds env0 binds0 = do { -- Put all the top-level binders into scope at the start -- so that if a transformation rule has unexpectedly brought -- anything into scope, then we don't get a complaint about that. -- It's rather as if the top-level binders were imported. -- See note [Glomming] in OccurAnal. ; env1 <- simplRecBndrs env0 (bindersOfBinds binds0) ; env2 <- simpl_binds env1 binds0 ; freeTick SimplifierDone ; return env2 } where -- We need to track the zapped top-level binders, because -- they should have their fragile IdInfo zapped (notably occurrence info) -- That's why we run down binds and bndrs' simultaneously. -- simpl_binds :: SimplEnv -> [InBind] -> SimplM SimplEnv simpl_binds env [] = return env simpl_binds env (bind:binds) = do { env' <- simpl_bind env bind ; simpl_binds env' binds } simpl_bind env (Rec pairs) = simplRecBind env TopLevel pairs simpl_bind env (NonRec b r) = do { (env', b') <- addBndrRules env b (lookupRecBndr env b) ; simplRecOrTopPair env' TopLevel NonRecursive b b' r } {- ************************************************************************ * * \subsection{Lazy bindings} * * ************************************************************************ simplRecBind is used for * recursive bindings only -} simplRecBind :: SimplEnv -> TopLevelFlag -> [(InId, InExpr)] -> SimplM SimplEnv simplRecBind env0 top_lvl pairs0 = do { (env_with_info, triples) <- mapAccumLM add_rules env0 pairs0 ; env1 <- go (zapFloats env_with_info) triples ; return (env0 `addRecFloats` env1) } -- addFloats adds the floats from env1, -- _and_ updates env0 with the in-scope set from env1 where add_rules :: SimplEnv -> (InBndr,InExpr) -> SimplM (SimplEnv, (InBndr, OutBndr, InExpr)) -- Add the (substituted) rules to the binder add_rules env (bndr, rhs) = do { (env', bndr') <- addBndrRules env bndr (lookupRecBndr env bndr) ; return (env', (bndr, bndr', rhs)) } go env [] = return env go env ((old_bndr, new_bndr, rhs) : pairs) = do { env' <- simplRecOrTopPair env top_lvl Recursive old_bndr new_bndr rhs ; go env' pairs } {- simplOrTopPair is used for * recursive bindings (whether top level or not) * top-level non-recursive bindings It assumes the binder has already been simplified, but not its IdInfo. -} simplRecOrTopPair :: SimplEnv -> TopLevelFlag -> RecFlag -> InId -> OutBndr -> InExpr -- Binder and rhs -> SimplM SimplEnv -- Returns an env that includes the binding simplRecOrTopPair env top_lvl is_rec old_bndr new_bndr rhs = do { dflags <- getDynFlags ; trace_bind dflags $ if preInlineUnconditionally dflags env top_lvl old_bndr rhs -- Check for unconditional inline then do tick (PreInlineUnconditionally old_bndr) return (extendIdSubst env old_bndr (mkContEx env rhs)) else simplLazyBind env top_lvl is_rec old_bndr new_bndr rhs env } where trace_bind dflags thing_inside | not (dopt Opt_D_verbose_core2core dflags) = thing_inside | otherwise = pprTrace "SimplBind" (ppr old_bndr) thing_inside -- trace_bind emits a trace for each top-level binding, which -- helps to locate the tracing for inlining and rule firing {- simplLazyBind is used for * [simplRecOrTopPair] recursive bindings (whether top level or not) * [simplRecOrTopPair] top-level non-recursive bindings * [simplNonRecE] non-top-level *lazy* non-recursive bindings Nota bene: 1. It assumes that the binder is *already* simplified, and is in scope, and its IdInfo too, except unfolding 2. It assumes that the binder type is lifted. 3. It does not check for pre-inline-unconditionally; that should have been done already. -} simplLazyBind :: SimplEnv -> TopLevelFlag -> RecFlag -> InId -> OutId -- Binder, both pre-and post simpl -- The OutId has IdInfo, except arity, unfolding -> InExpr -> SimplEnv -- The RHS and its environment -> SimplM SimplEnv -- Precondition: rhs obeys the let/app invariant simplLazyBind env top_lvl is_rec bndr bndr1 rhs rhs_se = -- pprTrace "simplLazyBind" ((ppr bndr <+> ppr bndr1) $$ ppr rhs $$ ppr (seIdSubst rhs_se)) $ do { let rhs_env = rhs_se `setInScope` env (tvs, body) = case collectTyBinders rhs of (tvs, body) | not_lam body -> (tvs,body) | otherwise -> ([], rhs) not_lam (Lam _ _) = False not_lam (Tick t e) | not (tickishFloatable t) = not_lam e -- eta-reduction could float not_lam _ = True -- Do not do the "abstract tyyvar" thing if there's -- a lambda inside, because it defeats eta-reduction -- f = /\a. \x. g a x -- should eta-reduce. ; (body_env, tvs') <- simplBinders rhs_env tvs -- See Note [Floating and type abstraction] in SimplUtils -- Simplify the RHS ; let rhs_cont = mkRhsStop (substTy body_env (exprType body)) ; (body_env1, body1) <- simplExprF body_env body rhs_cont -- ANF-ise a constructor or PAP rhs ; (body_env2, body2) <- prepareRhs top_lvl body_env1 bndr1 body1 ; (env', rhs') <- if not (doFloatFromRhs top_lvl is_rec False body2 body_env2) then -- No floating, revert to body1 do { rhs' <- mkLam tvs' (wrapFloats body_env1 body1) rhs_cont ; return (env, rhs') } else if null tvs then -- Simple floating do { tick LetFloatFromLet ; return (addFloats env body_env2, body2) } else -- Do type-abstraction first do { tick LetFloatFromLet ; (poly_binds, body3) <- abstractFloats tvs' body_env2 body2 ; rhs' <- mkLam tvs' body3 rhs_cont ; env' <- foldlM (addPolyBind top_lvl) env poly_binds ; return (env', rhs') } ; completeBind env' top_lvl bndr bndr1 rhs' } {- A specialised variant of simplNonRec used when the RHS is already simplified, notably in knownCon. It uses case-binding where necessary. -} simplNonRecX :: SimplEnv -> InId -- Old binder -> OutExpr -- Simplified RHS -> SimplM SimplEnv -- Precondition: rhs satisfies the let/app invariant simplNonRecX env bndr new_rhs | isDeadBinder bndr -- Not uncommon; e.g. case (a,b) of c { (p,q) -> p } = return env -- Here c is dead, and we avoid creating -- the binding c = (a,b) | Coercion co <- new_rhs = return (extendCvSubst env bndr co) | otherwise = do { (env', bndr') <- simplBinder env bndr ; completeNonRecX NotTopLevel env' (isStrictId bndr) bndr bndr' new_rhs } -- simplNonRecX is only used for NotTopLevel things completeNonRecX :: TopLevelFlag -> SimplEnv -> Bool -> InId -- Old binder -> OutId -- New binder -> OutExpr -- Simplified RHS -> SimplM SimplEnv -- Precondition: rhs satisfies the let/app invariant -- See Note [CoreSyn let/app invariant] in CoreSyn completeNonRecX top_lvl env is_strict old_bndr new_bndr new_rhs = do { (env1, rhs1) <- prepareRhs top_lvl (zapFloats env) new_bndr new_rhs ; (env2, rhs2) <- if doFloatFromRhs NotTopLevel NonRecursive is_strict rhs1 env1 then do { tick LetFloatFromLet ; return (addFloats env env1, rhs1) } -- Add the floats to the main env else return (env, wrapFloats env1 rhs1) -- Wrap the floats around the RHS ; completeBind env2 NotTopLevel old_bndr new_bndr rhs2 } {- {- No, no, no! Do not try preInlineUnconditionally in completeNonRecX Doing so risks exponential behaviour, because new_rhs has been simplified once already In the cases described by the folowing commment, postInlineUnconditionally will catch many of the relevant cases. -- This happens; for example, the case_bndr during case of -- known constructor: case (a,b) of x { (p,q) -> ... } -- Here x isn't mentioned in the RHS, so we don't want to -- create the (dead) let-binding let x = (a,b) in ... -- -- Similarly, single occurrences can be inlined vigourously -- e.g. case (f x, g y) of (a,b) -> .... -- If a,b occur once we can avoid constructing the let binding for them. Furthermore in the case-binding case preInlineUnconditionally risks extra thunks -- Consider case I# (quotInt# x y) of -- I# v -> let w = J# v in ... -- If we gaily inline (quotInt# x y) for v, we end up building an -- extra thunk: -- let w = J# (quotInt# x y) in ... -- because quotInt# can fail. | preInlineUnconditionally env NotTopLevel bndr new_rhs = thing_inside (extendIdSubst env bndr (DoneEx new_rhs)) -} ---------------------------------- prepareRhs takes a putative RHS, checks whether it's a PAP or constructor application and, if so, converts it to ANF, so that the resulting thing can be inlined more easily. Thus x = (f a, g b) becomes t1 = f a t2 = g b x = (t1,t2) We also want to deal well cases like this v = (f e1 `cast` co) e2 Here we want to make e1,e2 trivial and get x1 = e1; x2 = e2; v = (f x1 `cast` co) v2 That's what the 'go' loop in prepareRhs does -} prepareRhs :: TopLevelFlag -> SimplEnv -> OutId -> OutExpr -> SimplM (SimplEnv, OutExpr) -- Adds new floats to the env iff that allows us to return a good RHS prepareRhs top_lvl env id (Cast rhs co) -- Note [Float coercions] | Pair ty1 _ty2 <- coercionKind co -- Do *not* do this if rhs has an unlifted type , not (isUnLiftedType ty1) -- see Note [Float coercions (unlifted)] = do { (env', rhs') <- makeTrivialWithInfo top_lvl env sanitised_info rhs ; return (env', Cast rhs' co) } where sanitised_info = vanillaIdInfo `setStrictnessInfo` strictnessInfo info `setDemandInfo` demandInfo info info = idInfo id prepareRhs top_lvl env0 _ rhs0 = do { (_is_exp, env1, rhs1) <- go 0 env0 rhs0 ; return (env1, rhs1) } where go n_val_args env (Cast rhs co) = do { (is_exp, env', rhs') <- go n_val_args env rhs ; return (is_exp, env', Cast rhs' co) } go n_val_args env (App fun (Type ty)) = do { (is_exp, env', rhs') <- go n_val_args env fun ; return (is_exp, env', App rhs' (Type ty)) } go n_val_args env (App fun arg) = do { (is_exp, env', fun') <- go (n_val_args+1) env fun ; case is_exp of True -> do { (env'', arg') <- makeTrivial top_lvl env' arg ; return (True, env'', App fun' arg') } False -> return (False, env, App fun arg) } go n_val_args env (Var fun) = return (is_exp, env, Var fun) where is_exp = isExpandableApp fun n_val_args -- The fun a constructor or PAP -- See Note [CONLIKE pragma] in BasicTypes -- The definition of is_exp should match that in -- OccurAnal.occAnalApp go n_val_args env (Tick t rhs) -- We want to be able to float bindings past this -- tick. Non-scoping ticks don't care. | tickishScoped t == NoScope = do { (is_exp, env', rhs') <- go n_val_args env rhs ; return (is_exp, env', Tick t rhs') } -- On the other hand, for scoping ticks we need to be able to -- copy them on the floats, which in turn is only allowed if -- we can obtain non-counting ticks. | not (tickishCounts t) || tickishCanSplit t = do { (is_exp, env', rhs') <- go n_val_args (zapFloats env) rhs ; let tickIt (id, expr) = (id, mkTick (mkNoCount t) expr) floats' = seFloats $ env `addFloats` mapFloats env' tickIt ; return (is_exp, env' { seFloats = floats' }, Tick t rhs') } go _ env other = return (False, env, other) {- Note [Float coercions] ~~~~~~~~~~~~~~~~~~~~~~ When we find the binding x = e `cast` co we'd like to transform it to x' = e x = x `cast` co -- A trivial binding There's a chance that e will be a constructor application or function, or something like that, so moving the coercion to the usage site may well cancel the coercions and lead to further optimisation. Example: data family T a :: * data instance T Int = T Int foo :: Int -> Int -> Int foo m n = ... where x = T m go 0 = 0 go n = case x of { T m -> go (n-m) } -- This case should optimise Note [Preserve strictness when floating coercions] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In the Note [Float coercions] transformation, keep the strictness info. Eg f = e `cast` co -- f has strictness SSL When we transform to f' = e -- f' also has strictness SSL f = f' `cast` co -- f still has strictness SSL Its not wrong to drop it on the floor, but better to keep it. Note [Float coercions (unlifted)] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ BUT don't do [Float coercions] if 'e' has an unlifted type. This *can* happen: foo :: Int = (error (# Int,Int #) "urk") `cast` CoUnsafe (# Int,Int #) Int If do the makeTrivial thing to the error call, we'll get foo = case error (# Int,Int #) "urk" of v -> v `cast` ... But 'v' isn't in scope! These strange casts can happen as a result of case-of-case bar = case (case x of { T -> (# 2,3 #); F -> error "urk" }) of (# p,q #) -> p+q -} makeTrivialArg :: SimplEnv -> ArgSpec -> SimplM (SimplEnv, ArgSpec) makeTrivialArg env (ValArg e) = do { (env', e') <- makeTrivial NotTopLevel env e ; return (env', ValArg e') } makeTrivialArg env arg = return (env, arg) -- CastBy, TyArg makeTrivial :: TopLevelFlag -> SimplEnv -> OutExpr -> SimplM (SimplEnv, OutExpr) -- Binds the expression to a variable, if it's not trivial, returning the variable makeTrivial top_lvl env expr = makeTrivialWithInfo top_lvl env vanillaIdInfo expr makeTrivialWithInfo :: TopLevelFlag -> SimplEnv -> IdInfo -> OutExpr -> SimplM (SimplEnv, OutExpr) -- Propagate strictness and demand info to the new binder -- Note [Preserve strictness when floating coercions] -- Returned SimplEnv has same substitution as incoming one makeTrivialWithInfo top_lvl env info expr | exprIsTrivial expr -- Already trivial || not (bindingOk top_lvl expr expr_ty) -- Cannot trivialise -- See Note [Cannot trivialise] = return (env, expr) | otherwise -- See Note [Take care] below = do { uniq <- getUniqueM ; let name = mkSystemVarName uniq (fsLit "a") var = mkLocalIdWithInfo name expr_ty info ; env' <- completeNonRecX top_lvl env False var var expr ; expr' <- simplVar env' var ; return (env', expr') } -- The simplVar is needed becase we're constructing a new binding -- a = rhs -- And if rhs is of form (rhs1 |> co), then we might get -- a1 = rhs1 -- a = a1 |> co -- and now a's RHS is trivial and can be substituted out, and that -- is what completeNonRecX will do -- To put it another way, it's as if we'd simplified -- let var = e in var where expr_ty = exprType expr bindingOk :: TopLevelFlag -> CoreExpr -> Type -> Bool -- True iff we can have a binding of this expression at this level -- Precondition: the type is the type of the expression bindingOk top_lvl _ expr_ty | isTopLevel top_lvl = not (isUnLiftedType expr_ty) | otherwise = True {- Note [Cannot trivialise] ~~~~~~~~~~~~~~~~~~~~~~~~ Consider tih f :: Int -> Addr# foo :: Bar foo = Bar (f 3) Then we can't ANF-ise foo, even though we'd like to, because we can't make a top-level binding for the Addr# (f 3). And if so we don't want to turn it into foo = let x = f 3 in Bar x because we'll just end up inlining x back, and that makes the simplifier loop. Better not to ANF-ise it at all. A case in point is literal strings (a MachStr is not regarded as trivial): foo = Ptr "blob"# We don't want to ANF-ise this. ************************************************************************ * * \subsection{Completing a lazy binding} * * ************************************************************************ completeBind * deals only with Ids, not TyVars * takes an already-simplified binder and RHS * is used for both recursive and non-recursive bindings * is used for both top-level and non-top-level bindings It does the following: - tries discarding a dead binding - tries PostInlineUnconditionally - add unfolding [this is the only place we add an unfolding] - add arity It does *not* attempt to do let-to-case. Why? Because it is used for - top-level bindings (when let-to-case is impossible) - many situations where the "rhs" is known to be a WHNF (so let-to-case is inappropriate). Nor does it do the atomic-argument thing -} completeBind :: SimplEnv -> TopLevelFlag -- Flag stuck into unfolding -> InId -- Old binder -> OutId -> OutExpr -- New binder and RHS -> SimplM SimplEnv -- completeBind may choose to do its work -- * by extending the substitution (e.g. let x = y in ...) -- * or by adding to the floats in the envt -- -- Precondition: rhs obeys the let/app invariant completeBind env top_lvl old_bndr new_bndr new_rhs | isCoVar old_bndr = case new_rhs of Coercion co -> return (extendCvSubst env old_bndr co) _ -> return (addNonRec env new_bndr new_rhs) | otherwise = ASSERT( isId new_bndr ) do { let old_info = idInfo old_bndr old_unf = unfoldingInfo old_info occ_info = occInfo old_info -- Do eta-expansion on the RHS of the binding -- See Note [Eta-expanding at let bindings] in SimplUtils ; (new_arity, final_rhs) <- tryEtaExpandRhs env new_bndr new_rhs -- Simplify the unfolding ; new_unfolding <- simplLetUnfolding env top_lvl old_bndr final_rhs old_unf ; dflags <- getDynFlags ; if postInlineUnconditionally dflags env top_lvl new_bndr occ_info final_rhs new_unfolding -- Inline and discard the binding then do { tick (PostInlineUnconditionally old_bndr) ; return (extendIdSubst env old_bndr (DoneEx final_rhs)) } -- Use the substitution to make quite, quite sure that the -- substitution will happen, since we are going to discard the binding else do { let info1 = idInfo new_bndr `setArityInfo` new_arity -- Unfolding info: Note [Setting the new unfolding] info2 = info1 `setUnfoldingInfo` new_unfolding -- Demand info: Note [Setting the demand info] -- -- We also have to nuke demand info if for some reason -- eta-expansion *reduces* the arity of the binding to less -- than that of the strictness sig. This can happen: see Note [Arity decrease]. info3 | isEvaldUnfolding new_unfolding || (case strictnessInfo info2 of StrictSig dmd_ty -> new_arity < dmdTypeDepth dmd_ty) = zapDemandInfo info2 `orElse` info2 | otherwise = info2 final_id = new_bndr `setIdInfo` info3 ; -- pprTrace "Binding" (ppr final_id <+> ppr new_unfolding) $ return (addNonRec env final_id final_rhs) } } -- The addNonRec adds it to the in-scope set too ------------------------------ addPolyBind :: TopLevelFlag -> SimplEnv -> OutBind -> SimplM SimplEnv -- Add a new binding to the environment, complete with its unfolding -- but *do not* do postInlineUnconditionally, because we have already -- processed some of the scope of the binding -- We still want the unfolding though. Consider -- let -- x = /\a. let y = ... in Just y -- in body -- Then we float the y-binding out (via abstractFloats and addPolyBind) -- but 'x' may well then be inlined in 'body' in which case we'd like the -- opportunity to inline 'y' too. -- -- INVARIANT: the arity is correct on the incoming binders addPolyBind top_lvl env (NonRec poly_id rhs) = do { unfolding <- simplLetUnfolding env top_lvl poly_id rhs noUnfolding -- Assumes that poly_id did not have an INLINE prag -- which is perhaps wrong. ToDo: think about this ; let final_id = setIdInfo poly_id $ idInfo poly_id `setUnfoldingInfo` unfolding ; return (addNonRec env final_id rhs) } addPolyBind _ env bind@(Rec _) = return (extendFloats env bind) -- Hack: letrecs are more awkward, so we extend "by steam" -- without adding unfoldings etc. At worst this leads to -- more simplifier iterations {- Note [Arity decrease] ~~~~~~~~~~~~~~~~~~~~~~~~ Generally speaking the arity of a binding should not decrease. But it *can* legitimately happen because of RULES. Eg f = g Int where g has arity 2, will have arity 2. But if there's a rewrite rule g Int --> h where h has arity 1, then f's arity will decrease. Here's a real-life example, which is in the output of Specialise: Rec { $dm {Arity 2} = \d.\x. op d {-# RULES forall d. $dm Int d = $s$dm #-} dInt = MkD .... opInt ... opInt {Arity 1} = $dm dInt $s$dm {Arity 0} = \x. op dInt } Here opInt has arity 1; but when we apply the rule its arity drops to 0. That's why Specialise goes to a little trouble to pin the right arity on specialised functions too. Note [Setting the demand info] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ If the unfolding is a value, the demand info may go pear-shaped, so we nuke it. Example: let x = (a,b) in case x of (p,q) -> h p q x Here x is certainly demanded. But after we've nuked the case, we'll get just let x = (a,b) in h a b x and now x is not demanded (I'm assuming h is lazy) This really happens. Similarly let f = \x -> e in ...f..f... After inlining f at some of its call sites the original binding may (for example) be no longer strictly demanded. The solution here is a bit ad hoc... ************************************************************************ * * \subsection[Simplify-simplExpr]{The main function: simplExpr} * * ************************************************************************ The reason for this OutExprStuff stuff is that we want to float *after* simplifying a RHS, not before. If we do so naively we get quadratic behaviour as things float out. To see why it's important to do it after, consider this (real) example: let t = f x in fst t ==> let t = let a = e1 b = e2 in (a,b) in fst t ==> let a = e1 b = e2 t = (a,b) in a -- Can't inline a this round, cos it appears twice ==> e1 Each of the ==> steps is a round of simplification. We'd save a whole round if we float first. This can cascade. Consider let f = g d in \x -> ...f... ==> let f = let d1 = ..d.. in \y -> e in \x -> ...f... ==> let d1 = ..d.. in \x -> ...(\y ->e)... Only in this second round can the \y be applied, and it might do the same again. -} simplExpr :: SimplEnv -> CoreExpr -> SimplM CoreExpr simplExpr env expr = simplExprC env expr (mkBoringStop expr_out_ty) where expr_out_ty :: OutType expr_out_ty = substTy env (exprType expr) simplExprC :: SimplEnv -> CoreExpr -> SimplCont -> SimplM CoreExpr -- Simplify an expression, given a continuation simplExprC env expr cont = -- pprTrace "simplExprC" (ppr expr $$ ppr cont {- $$ ppr (seIdSubst env) -} $$ ppr (seFloats env) ) $ do { (env', expr') <- simplExprF (zapFloats env) expr cont ; -- pprTrace "simplExprC ret" (ppr expr $$ ppr expr') $ -- pprTrace "simplExprC ret3" (ppr (seInScope env')) $ -- pprTrace "simplExprC ret4" (ppr (seFloats env')) $ return (wrapFloats env' expr') } -------------------------------------------------- simplExprF :: SimplEnv -> InExpr -> SimplCont -> SimplM (SimplEnv, OutExpr) simplExprF env e cont = {- pprTrace "simplExprF" (vcat [ ppr e , text "cont =" <+> ppr cont , text "inscope =" <+> ppr (seInScope env) , text "tvsubst =" <+> ppr (seTvSubst env) , text "idsubst =" <+> ppr (seIdSubst env) , text "cvsubst =" <+> ppr (seCvSubst env) {- , ppr (seFloats env) -} ]) $ -} simplExprF1 env e cont simplExprF1 :: SimplEnv -> InExpr -> SimplCont -> SimplM (SimplEnv, OutExpr) simplExprF1 env (Var v) cont = simplIdF env v cont simplExprF1 env (Lit lit) cont = rebuild env (Lit lit) cont simplExprF1 env (Tick t expr) cont = simplTick env t expr cont simplExprF1 env (Cast body co) cont = simplCast env body co cont simplExprF1 env (Coercion co) cont = simplCoercionF env co cont simplExprF1 env (Type ty) cont = ASSERT( contIsRhsOrArg cont ) rebuild env (Type (substTy env ty)) cont simplExprF1 env (App fun arg) cont = simplExprF env fun $ case arg of Type ty -> ApplyToTy { sc_arg_ty = substTy env ty , sc_hole_ty = substTy env (exprType fun) , sc_cont = cont } _ -> ApplyToVal { sc_arg = arg, sc_env = env , sc_dup = NoDup, sc_cont = cont } simplExprF1 env expr@(Lam {}) cont = simplLam env zapped_bndrs body cont -- The main issue here is under-saturated lambdas -- (\x1. \x2. e) arg1 -- Here x1 might have "occurs-once" occ-info, because occ-info -- is computed assuming that a group of lambdas is applied -- all at once. If there are too few args, we must zap the -- occ-info, UNLESS the remaining binders are one-shot where (bndrs, body) = collectBinders expr zapped_bndrs | need_to_zap = map zap bndrs | otherwise = bndrs need_to_zap = any zappable_bndr (drop n_args bndrs) n_args = countArgs cont -- NB: countArgs counts all the args (incl type args) -- and likewise drop counts all binders (incl type lambdas) zappable_bndr b = isId b && not (isOneShotBndr b) zap b | isTyVar b = b | otherwise = zapLamIdInfo b simplExprF1 env (Case scrut bndr _ alts) cont = simplExprF env scrut (Select { sc_dup = NoDup, sc_bndr = bndr , sc_alts = alts , sc_env = env, sc_cont = cont }) simplExprF1 env (Let (Rec pairs) body) cont = do { env' <- simplRecBndrs env (map fst pairs) -- NB: bndrs' don't have unfoldings or rules -- We add them as we go down ; env'' <- simplRecBind env' NotTopLevel pairs ; simplExprF env'' body cont } simplExprF1 env (Let (NonRec bndr rhs) body) cont = simplNonRecE env bndr (rhs, env) ([], body) cont --------------------------------- simplType :: SimplEnv -> InType -> SimplM OutType -- Kept monadic just so we can do the seqType simplType env ty = -- pprTrace "simplType" (ppr ty $$ ppr (seTvSubst env)) $ seqType new_ty `seq` return new_ty where new_ty = substTy env ty --------------------------------- simplCoercionF :: SimplEnv -> InCoercion -> SimplCont -> SimplM (SimplEnv, OutExpr) simplCoercionF env co cont = do { co' <- simplCoercion env co ; rebuild env (Coercion co') cont } simplCoercion :: SimplEnv -> InCoercion -> SimplM OutCoercion simplCoercion env co = let opt_co = optCoercion (getCvSubst env) co in seqCo opt_co `seq` return opt_co ----------------------------------- -- | Push a TickIt context outwards past applications and cases, as -- long as this is a non-scoping tick, to let case and application -- optimisations apply. simplTick :: SimplEnv -> Tickish Id -> InExpr -> SimplCont -> SimplM (SimplEnv, OutExpr) simplTick env tickish expr cont -- A scoped tick turns into a continuation, so that we can spot -- (scc t (\x . e)) in simplLam and eliminate the scc. If we didn't do -- it this way, then it would take two passes of the simplifier to -- reduce ((scc t (\x . e)) e'). -- NB, don't do this with counting ticks, because if the expr is -- bottom, then rebuildCall will discard the continuation. -- XXX: we cannot do this, because the simplifier assumes that -- the context can be pushed into a case with a single branch. e.g. -- scc<f> case expensive of p -> e -- becomes -- case expensive of p -> scc<f> e -- -- So I'm disabling this for now. It just means we will do more -- simplifier iterations that necessary in some cases. -- | tickishScoped tickish && not (tickishCounts tickish) -- = simplExprF env expr (TickIt tickish cont) -- For unscoped or soft-scoped ticks, we are allowed to float in new -- cost, so we simply push the continuation inside the tick. This -- has the effect of moving the tick to the outside of a case or -- application context, allowing the normal case and application -- optimisations to fire. | tickish `tickishScopesLike` SoftScope = do { (env', expr') <- simplExprF env expr cont ; return (env', mkTick tickish expr') } -- Push tick inside if the context looks like this will allow us to -- do a case-of-case - see Note [case-of-scc-of-case] | Select {} <- cont, Just expr' <- push_tick_inside = simplExprF env expr' cont -- We don't want to move the tick, but we might still want to allow -- floats to pass through with appropriate wrapping (or not, see -- wrap_floats below) --- | not (tickishCounts tickish) || tickishCanSplit tickish -- = wrap_floats | otherwise = no_floating_past_tick where -- Try to push tick inside a case, see Note [case-of-scc-of-case]. push_tick_inside = case expr0 of Case scrut bndr ty alts -> Just $ Case (tickScrut scrut) bndr ty (map tickAlt alts) _other -> Nothing where (ticks, expr0) = stripTicksTop movable (Tick tickish expr) movable t = not (tickishCounts t) || t `tickishScopesLike` NoScope || tickishCanSplit t tickScrut e = foldr mkTick e ticks -- Alternatives get annotated with all ticks that scope in some way, -- but we don't want to count entries. tickAlt (c,bs,e) = (c,bs, foldr mkTick e ts_scope) ts_scope = map mkNoCount $ filter (not . (`tickishScopesLike` NoScope)) ticks no_floating_past_tick = do { let (inc,outc) = splitCont cont ; (env', expr') <- simplExprF (zapFloats env) expr inc ; let tickish' = simplTickish env tickish ; (env'', expr'') <- rebuild (zapFloats env') (wrapFloats env' expr') (TickIt tickish' outc) ; return (addFloats env env'', expr'') } -- Alternative version that wraps outgoing floats with the tick. This -- results in ticks being duplicated, as we don't make any attempt to -- eliminate the tick if we re-inline the binding (because the tick -- semantics allows unrestricted inlining of HNFs), so I'm not doing -- this any more. FloatOut will catch any real opportunities for -- floating. -- -- wrap_floats = -- do { let (inc,outc) = splitCont cont -- ; (env', expr') <- simplExprF (zapFloats env) expr inc -- ; let tickish' = simplTickish env tickish -- ; let wrap_float (b,rhs) = (zapIdStrictness (setIdArity b 0), -- mkTick (mkNoCount tickish') rhs) -- -- when wrapping a float with mkTick, we better zap the Id's -- -- strictness info and arity, because it might be wrong now. -- ; let env'' = addFloats env (mapFloats env' wrap_float) -- ; rebuild env'' expr' (TickIt tickish' outc) -- } simplTickish env tickish | Breakpoint n ids <- tickish = Breakpoint n (map (getDoneId . substId env) ids) | otherwise = tickish -- Push type application and coercion inside a tick splitCont :: SimplCont -> (SimplCont, SimplCont) splitCont cont@(ApplyToTy { sc_cont = tail }) = (cont { sc_cont = inc }, outc) where (inc,outc) = splitCont tail splitCont (CastIt co c) = (CastIt co inc, outc) where (inc,outc) = splitCont c splitCont other = (mkBoringStop (contHoleType other), other) getDoneId (DoneId id) = id getDoneId (DoneEx e) = getIdFromTrivialExpr e -- Note [substTickish] in CoreSubst getDoneId other = pprPanic "getDoneId" (ppr other) -- Note [case-of-scc-of-case] -- It's pretty important to be able to transform case-of-case when -- there's an SCC in the way. For example, the following comes up -- in nofib/real/compress/Encode.hs: -- -- case scctick<code_string.r1> -- case $wcode_string_r13s wild_XC w1_s137 w2_s138 l_aje -- of _ { (# ww1_s13f, ww2_s13g, ww3_s13h #) -> -- (ww1_s13f, ww2_s13g, ww3_s13h) -- } -- of _ { (ww_s12Y, ww1_s12Z, ww2_s130) -> -- tick<code_string.f1> -- (ww_s12Y, -- ww1_s12Z, -- PTTrees.PT -- @ GHC.Types.Char @ GHC.Types.Int wild2_Xj ww2_s130 r_ajf) -- } -- -- We really want this case-of-case to fire, because then the 3-tuple -- will go away (indeed, the CPR optimisation is relying on this -- happening). But the scctick is in the way - we need to push it -- inside to expose the case-of-case. So we perform this -- transformation on the inner case: -- -- scctick c (case e of { p1 -> e1; ...; pn -> en }) -- ==> -- case (scctick c e) of { p1 -> scc c e1; ...; pn -> scc c en } -- -- So we've moved a constant amount of work out of the scc to expose -- the case. We only do this when the continuation is interesting: in -- for now, it has to be another Case (maybe generalise this later). {- ************************************************************************ * * \subsection{The main rebuilder} * * ************************************************************************ -} rebuild :: SimplEnv -> OutExpr -> SimplCont -> SimplM (SimplEnv, OutExpr) -- At this point the substitution in the SimplEnv should be irrelevant -- only the in-scope set and floats should matter rebuild env expr cont = case cont of Stop {} -> return (env, expr) TickIt t cont -> rebuild env (mkTick t expr) cont CastIt co cont -> rebuild env (mkCast expr co) cont -- NB: mkCast implements the (Coercion co |> g) optimisation Select { sc_bndr = bndr, sc_alts = alts, sc_env = se, sc_cont = cont } -> rebuildCase (se `setFloats` env) expr bndr alts cont StrictArg info _ cont -> rebuildCall env (info `addValArgTo` expr) cont StrictBind b bs body se cont -> do { env' <- simplNonRecX (se `setFloats` env) b expr -- expr satisfies let/app since it started life -- in a call to simplNonRecE ; simplLam env' bs body cont } ApplyToTy { sc_arg_ty = ty, sc_cont = cont} -> rebuild env (App expr (Type ty)) cont ApplyToVal { sc_arg = arg, sc_env = se, sc_dup = dup_flag, sc_cont = cont} -- See Note [Avoid redundant simplification] | isSimplified dup_flag -> rebuild env (App expr arg) cont | otherwise -> do { arg' <- simplExpr (se `setInScope` env) arg ; rebuild env (App expr arg') cont } {- ************************************************************************ * * \subsection{Lambdas} * * ************************************************************************ -} simplCast :: SimplEnv -> InExpr -> Coercion -> SimplCont -> SimplM (SimplEnv, OutExpr) simplCast env body co0 cont0 = do { co1 <- simplCoercion env co0 ; cont1 <- addCoerce co1 cont0 ; simplExprF env body cont1 } where addCoerce co cont = add_coerce co (coercionKind co) cont add_coerce _co (Pair s1 k1) cont -- co :: ty~ty | s1 `eqType` k1 = return cont -- is a no-op add_coerce co1 (Pair s1 _k2) (CastIt co2 cont) | (Pair _l1 t1) <- coercionKind co2 -- e |> (g1 :: S1~L) |> (g2 :: L~T1) -- ==> -- e, if S1=T1 -- e |> (g1 . g2 :: S1~T1) otherwise -- -- For example, in the initial form of a worker -- we may find (coerce T (coerce S (\x.e))) y -- and we'd like it to simplify to e[y/x] in one round -- of simplification , s1 `eqType` t1 = return cont -- The coerces cancel out | otherwise = return (CastIt (mkTransCo co1 co2) cont) add_coerce co (Pair s1s2 _t1t2) cont@(ApplyToTy { sc_arg_ty = arg_ty, sc_cont = tail }) -- (f |> g) ty ---> (f ty) |> (g @ ty) -- This implements the PushT rule from the paper | Just (tyvar,_) <- splitForAllTy_maybe s1s2 = ASSERT( isTyVar tyvar ) do { cont' <- addCoerce new_cast tail ; return (cont { sc_cont = cont' }) } where new_cast = mkInstCo co arg_ty add_coerce co (Pair s1s2 t1t2) (ApplyToVal { sc_arg = arg, sc_env = arg_se , sc_dup = dup, sc_cont = cont }) | isFunTy s1s2 -- This implements the Push rule from the paper , isFunTy t1t2 -- Check t1t2 to ensure 'arg' is a value arg -- (e |> (g :: s1s2 ~ t1->t2)) f -- ===> -- (e (f |> (arg g :: t1~s1)) -- |> (res g :: s2->t2) -- -- t1t2 must be a function type, t1->t2, because it's applied -- to something but s1s2 might conceivably not be -- -- When we build the ApplyTo we can't mix the out-types -- with the InExpr in the argument, so we simply substitute -- to make it all consistent. It's a bit messy. -- But it isn't a common case. -- -- Example of use: Trac #995 = do { (dup', arg_se', arg') <- simplArg env dup arg_se arg ; cont' <- addCoerce co2 cont ; return (ApplyToVal { sc_arg = mkCast arg' (mkSymCo co1) , sc_env = arg_se' , sc_dup = dup' , sc_cont = cont' }) } where -- we split coercion t1->t2 ~ s1->s2 into t1 ~ s1 and -- t2 ~ s2 with left and right on the curried form: -- (->) t1 t2 ~ (->) s1 s2 [co1, co2] = decomposeCo 2 co add_coerce co _ cont = return (CastIt co cont) simplArg :: SimplEnv -> DupFlag -> StaticEnv -> CoreExpr -> SimplM (DupFlag, StaticEnv, OutExpr) simplArg env dup_flag arg_env arg | isSimplified dup_flag = return (dup_flag, arg_env, arg) | otherwise = do { arg' <- simplExpr (arg_env `setInScope` env) arg ; return (Simplified, zapSubstEnv arg_env, arg') } {- ************************************************************************ * * \subsection{Lambdas} * * ************************************************************************ Note [Zap unfolding when beta-reducing] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Lambda-bound variables can have stable unfoldings, such as $j = \x. \b{Unf=Just x}. e See Note [Case binders and join points] below; the unfolding for lets us optimise e better. However when we beta-reduce it we want to revert to using the actual value, otherwise we can end up in the stupid situation of let x = blah in let b{Unf=Just x} = y in ...b... Here it'd be far better to drop the unfolding and use the actual RHS. -} simplLam :: SimplEnv -> [InId] -> InExpr -> SimplCont -> SimplM (SimplEnv, OutExpr) simplLam env [] body cont = simplExprF env body cont -- Beta reduction simplLam env (bndr:bndrs) body (ApplyToTy { sc_arg_ty = arg_ty, sc_cont = cont }) = do { tick (BetaReduction bndr) ; simplLam (extendTvSubst env bndr arg_ty) bndrs body cont } simplLam env (bndr:bndrs) body (ApplyToVal { sc_arg = arg, sc_env = arg_se , sc_cont = cont }) = do { tick (BetaReduction bndr) ; simplNonRecE env (zap_unfolding bndr) (arg, arg_se) (bndrs, body) cont } where zap_unfolding bndr -- See Note [Zap unfolding when beta-reducing] | isId bndr, isStableUnfolding (realIdUnfolding bndr) = setIdUnfolding bndr NoUnfolding | otherwise = bndr -- discard a non-counting tick on a lambda. This may change the -- cost attribution slightly (moving the allocation of the -- lambda elsewhere), but we don't care: optimisation changes -- cost attribution all the time. simplLam env bndrs body (TickIt tickish cont) | not (tickishCounts tickish) = simplLam env bndrs body cont -- Not enough args, so there are real lambdas left to put in the result simplLam env bndrs body cont = do { (env', bndrs') <- simplLamBndrs env bndrs ; body' <- simplExpr env' body ; new_lam <- mkLam bndrs' body' cont ; rebuild env' new_lam cont } simplLamBndrs :: SimplEnv -> [InBndr] -> SimplM (SimplEnv, [OutBndr]) simplLamBndrs env bndrs = mapAccumLM simplLamBndr env bndrs ------------- simplLamBndr :: SimplEnv -> Var -> SimplM (SimplEnv, Var) -- Used for lambda binders. These sometimes have unfoldings added by -- the worker/wrapper pass that must be preserved, because they can't -- be reconstructed from context. For example: -- f x = case x of (a,b) -> fw a b x -- fw a b x{=(a,b)} = ... -- The "{=(a,b)}" is an unfolding we can't reconstruct otherwise. simplLamBndr env bndr | isId bndr && hasSomeUnfolding old_unf -- Special case = do { (env1, bndr1) <- simplBinder env bndr ; unf' <- simplUnfolding env1 NotTopLevel bndr old_unf ; let bndr2 = bndr1 `setIdUnfolding` unf' ; return (modifyInScope env1 bndr2, bndr2) } | otherwise = simplBinder env bndr -- Normal case where old_unf = idUnfolding bndr ------------------ simplNonRecE :: SimplEnv -> InBndr -- The binder -> (InExpr, SimplEnv) -- Rhs of binding (or arg of lambda) -> ([InBndr], InExpr) -- Body of the let/lambda -- \xs.e -> SimplCont -> SimplM (SimplEnv, OutExpr) -- simplNonRecE is used for -- * non-top-level non-recursive lets in expressions -- * beta reduction -- -- It deals with strict bindings, via the StrictBind continuation, -- which may abort the whole process -- -- Precondition: rhs satisfies the let/app invariant -- Note [CoreSyn let/app invariant] in CoreSyn -- -- The "body" of the binding comes as a pair of ([InId],InExpr) -- representing a lambda; so we recurse back to simplLam -- Why? Because of the binder-occ-info-zapping done before -- the call to simplLam in simplExprF (Lam ...) -- First deal with type applications and type lets -- (/\a. e) (Type ty) and (let a = Type ty in e) simplNonRecE env bndr (Type ty_arg, rhs_se) (bndrs, body) cont = ASSERT( isTyVar bndr ) do { ty_arg' <- simplType (rhs_se `setInScope` env) ty_arg ; simplLam (extendTvSubst env bndr ty_arg') bndrs body cont } simplNonRecE env bndr (rhs, rhs_se) (bndrs, body) cont = do dflags <- getDynFlags case () of _ | preInlineUnconditionally dflags env NotTopLevel bndr rhs -> do { tick (PreInlineUnconditionally bndr) ; -- pprTrace "preInlineUncond" (ppr bndr <+> ppr rhs) $ simplLam (extendIdSubst env bndr (mkContEx rhs_se rhs)) bndrs body cont } | isStrictId bndr -- Includes coercions -> simplExprF (rhs_se `setFloats` env) rhs (StrictBind bndr bndrs body env cont) | otherwise -> ASSERT( not (isTyVar bndr) ) do { (env1, bndr1) <- simplNonRecBndr env bndr ; (env2, bndr2) <- addBndrRules env1 bndr bndr1 ; env3 <- simplLazyBind env2 NotTopLevel NonRecursive bndr bndr2 rhs rhs_se ; simplLam env3 bndrs body cont } {- ************************************************************************ * * Variables * * ************************************************************************ -} simplVar :: SimplEnv -> InVar -> SimplM OutExpr -- Look up an InVar in the environment simplVar env var | isTyVar var = return (Type (substTyVar env var)) | isCoVar var = return (Coercion (substCoVar env var)) | otherwise = case substId env var of DoneId var1 -> return (Var var1) DoneEx e -> return e ContEx tvs cvs ids e -> simplExpr (setSubstEnv env tvs cvs ids) e simplIdF :: SimplEnv -> InId -> SimplCont -> SimplM (SimplEnv, OutExpr) simplIdF env var cont = case substId env var of DoneEx e -> simplExprF (zapSubstEnv env) e cont ContEx tvs cvs ids e -> simplExprF (setSubstEnv env tvs cvs ids) e cont DoneId var1 -> completeCall env var1 cont -- Note [zapSubstEnv] -- The template is already simplified, so don't re-substitute. -- This is VITAL. Consider -- let x = e in -- let y = \z -> ...x... in -- \ x -> ...y... -- We'll clone the inner \x, adding x->x' in the id_subst -- Then when we inline y, we must *not* replace x by x' in -- the inlined copy!! --------------------------------------------------------- -- Dealing with a call site completeCall :: SimplEnv -> OutId -> SimplCont -> SimplM (SimplEnv, OutExpr) completeCall env var cont = do { ------------- Try inlining ---------------- dflags <- getDynFlags ; let (lone_variable, arg_infos, call_cont) = contArgs cont n_val_args = length arg_infos interesting_cont = interestingCallContext call_cont unfolding = activeUnfolding env var maybe_inline = callSiteInline dflags var unfolding lone_variable arg_infos interesting_cont ; case maybe_inline of { Just expr -- There is an inlining! -> do { checkedTick (UnfoldingDone var) ; dump_inline dflags expr cont ; simplExprF (zapSubstEnv env) expr cont } ; Nothing -> do -- No inlining! { rule_base <- getSimplRules ; let info = mkArgInfo var (getRules rule_base var) n_val_args call_cont ; rebuildCall env info cont }}} where dump_inline dflags unfolding cont | not (dopt Opt_D_dump_inlinings dflags) = return () | not (dopt Opt_D_verbose_core2core dflags) = when (isExternalName (idName var)) $ liftIO $ printOutputForUser dflags alwaysQualify $ sep [text "Inlining done:", nest 4 (ppr var)] | otherwise = liftIO $ printOutputForUser dflags alwaysQualify $ sep [text "Inlining done: " <> ppr var, nest 4 (vcat [text "Inlined fn: " <+> nest 2 (ppr unfolding), text "Cont: " <+> ppr cont])] rebuildCall :: SimplEnv -> ArgInfo -> SimplCont -> SimplM (SimplEnv, OutExpr) rebuildCall env (ArgInfo { ai_fun = fun, ai_args = rev_args, ai_strs = [] }) cont -- When we run out of strictness args, it means -- that the call is definitely bottom; see SimplUtils.mkArgInfo -- Then we want to discard the entire strict continuation. E.g. -- * case (error "hello") of { ... } -- * (error "Hello") arg -- * f (error "Hello") where f is strict -- etc -- Then, especially in the first of these cases, we'd like to discard -- the continuation, leaving just the bottoming expression. But the -- type might not be right, so we may have to add a coerce. | not (contIsTrivial cont) -- Only do this if there is a non-trivial = return (env, castBottomExpr res cont_ty) -- contination to discard, else we do it where -- again and again! res = argInfoExpr fun rev_args cont_ty = contResultType cont rebuildCall env info (CastIt co cont) = rebuildCall env (addCastTo info co) cont rebuildCall env info (ApplyToTy { sc_arg_ty = arg_ty, sc_cont = cont }) = rebuildCall env (info `addTyArgTo` arg_ty) cont rebuildCall env info@(ArgInfo { ai_encl = encl_rules, ai_type = fun_ty , ai_strs = str:strs, ai_discs = disc:discs }) (ApplyToVal { sc_arg = arg, sc_env = arg_se , sc_dup = dup_flag, sc_cont = cont }) | isSimplified dup_flag -- See Note [Avoid redundant simplification] = rebuildCall env (addValArgTo info' arg) cont | str -- Strict argument = -- pprTrace "Strict Arg" (ppr arg $$ ppr (seIdSubst env) $$ ppr (seInScope env)) $ simplExprF (arg_se `setFloats` env) arg (StrictArg info' cci cont) -- Note [Shadowing] | otherwise -- Lazy argument -- DO NOT float anything outside, hence simplExprC -- There is no benefit (unlike in a let-binding), and we'd -- have to be very careful about bogus strictness through -- floating a demanded let. = do { arg' <- simplExprC (arg_se `setInScope` env) arg (mkLazyArgStop (funArgTy fun_ty) cci) ; rebuildCall env (addValArgTo info' arg') cont } where info' = info { ai_strs = strs, ai_discs = discs } cci | encl_rules = RuleArgCtxt | disc > 0 = DiscArgCtxt -- Be keener here | otherwise = BoringCtxt -- Nothing interesting rebuildCall env (ArgInfo { ai_fun = fun, ai_args = rev_args, ai_rules = rules }) cont | null rules = rebuild env (argInfoExpr fun rev_args) cont -- No rules, common case | otherwise = do { -- We've accumulated a simplified call in <fun,rev_args> -- so try rewrite rules; see Note [RULEs apply to simplified arguments] -- See also Note [Rules for recursive functions] ; let env' = zapSubstEnv env -- See Note [zapSubstEnv]; -- and NB that 'rev_args' are all fully simplified ; mb_rule <- tryRules env' rules fun (reverse rev_args) cont ; case mb_rule of { Just (rule_rhs, cont') -> simplExprF env' rule_rhs cont' -- Rules don't match ; Nothing -> rebuild env (argInfoExpr fun rev_args) cont -- No rules } } {- Note [RULES apply to simplified arguments] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ It's very desirable to try RULES once the arguments have been simplified, because doing so ensures that rule cascades work in one pass. Consider {-# RULES g (h x) = k x f (k x) = x #-} ...f (g (h x))... Then we want to rewrite (g (h x)) to (k x) and only then try f's rules. If we match f's rules against the un-simplified RHS, it won't match. This makes a particularly big difference when superclass selectors are involved: op ($p1 ($p2 (df d))) We want all this to unravel in one sweeep. Note [Avoid redundant simplification] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Because RULES apply to simplified arguments, there's a danger of repeatedly simplifying already-simplified arguments. An important example is that of (>>=) d e1 e2 Here e1, e2 are simplified before the rule is applied, but don't really participate in the rule firing. So we mark them as Simplified to avoid re-simplifying them. Note [Shadowing] ~~~~~~~~~~~~~~~~ This part of the simplifier may break the no-shadowing invariant Consider f (...(\a -> e)...) (case y of (a,b) -> e') where f is strict in its second arg If we simplify the innermost one first we get (...(\a -> e)...) Simplifying the second arg makes us float the case out, so we end up with case y of (a,b) -> f (...(\a -> e)...) e' So the output does not have the no-shadowing invariant. However, there is no danger of getting name-capture, because when the first arg was simplified we used an in-scope set that at least mentioned all the variables free in its static environment, and that is enough. We can't just do innermost first, or we'd end up with a dual problem: case x of (a,b) -> f e (...(\a -> e')...) I spent hours trying to recover the no-shadowing invariant, but I just could not think of an elegant way to do it. The simplifier is already knee-deep in continuations. We have to keep the right in-scope set around; AND we have to get the effect that finding (error "foo") in a strict arg position will discard the entire application and replace it with (error "foo"). Getting all this at once is TOO HARD! ************************************************************************ * * Rewrite rules * * ************************************************************************ -} tryRules :: SimplEnv -> [CoreRule] -> Id -> [ArgSpec] -> SimplCont -> SimplM (Maybe (CoreExpr, SimplCont)) -- The SimplEnv already has zapSubstEnv applied to it tryRules env rules fn args call_cont | null rules = return Nothing {- Disabled until we fix #8326 | fn `hasKey` tagToEnumKey -- See Note [Optimising tagToEnum#] , [_type_arg, val_arg] <- args , Select dup bndr ((_,[],rhs1) : rest_alts) se cont <- call_cont , isDeadBinder bndr = do { dflags <- getDynFlags ; let enum_to_tag :: CoreAlt -> CoreAlt -- Takes K -> e into tagK# -> e -- where tagK# is the tag of constructor K enum_to_tag (DataAlt con, [], rhs) = ASSERT( isEnumerationTyCon (dataConTyCon con) ) (LitAlt tag, [], rhs) where tag = mkMachInt dflags (toInteger (dataConTag con - fIRST_TAG)) enum_to_tag alt = pprPanic "tryRules: tagToEnum" (ppr alt) new_alts = (DEFAULT, [], rhs1) : map enum_to_tag rest_alts new_bndr = setIdType bndr intPrimTy -- The binder is dead, but should have the right type ; return (Just (val_arg, Select dup new_bndr new_alts se cont)) } -} | otherwise = do { dflags <- getDynFlags ; case lookupRule dflags (getUnfoldingInRuleMatch env) (activeRule env) fn (argInfoAppArgs args) rules of { Nothing -> return Nothing ; -- No rule matches Just (rule, rule_rhs) -> do { checkedTick (RuleFired (ru_name rule)) ; let cont' = pushSimplifiedArgs env (drop (ruleArity rule) args) call_cont -- (ruleArity rule) says how many args the rule consumed ; dump dflags rule rule_rhs ; return (Just (rule_rhs, cont')) }}} where dump dflags rule rule_rhs | dopt Opt_D_dump_rule_rewrites dflags = log_rule dflags Opt_D_dump_rule_rewrites "Rule fired" $ vcat [ text "Rule:" <+> ftext (ru_name rule) , text "Before:" <+> hang (ppr fn) 2 (sep (map ppr args)) , text "After: " <+> pprCoreExpr rule_rhs , text "Cont: " <+> ppr call_cont ] | dopt Opt_D_dump_rule_firings dflags = log_rule dflags Opt_D_dump_rule_firings "Rule fired:" $ ftext (ru_name rule) | otherwise = return () log_rule dflags flag hdr details = liftIO . dumpSDoc dflags alwaysQualify flag "" $ sep [text hdr, nest 4 details] {- Note [Optimising tagToEnum#] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~ If we have an enumeration data type: data Foo = A | B | C Then we want to transform case tagToEnum# x of ==> case x of A -> e1 DEFAULT -> e1 B -> e2 1# -> e2 C -> e3 2# -> e3 thereby getting rid of the tagToEnum# altogether. If there was a DEFAULT alternative we retain it (remember it comes first). If not the case must be exhaustive, and we reflect that in the transformed version by adding a DEFAULT. Otherwise Lint complains that the new case is not exhaustive. See #8317. Note [Rules for recursive functions] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ You might think that we shouldn't apply rules for a loop breaker: doing so might give rise to an infinite loop, because a RULE is rather like an extra equation for the function: RULE: f (g x) y = x+y Eqn: f a y = a-y But it's too drastic to disable rules for loop breakers. Even the foldr/build rule would be disabled, because foldr is recursive, and hence a loop breaker: foldr k z (build g) = g k z So it's up to the programmer: rules can cause divergence ************************************************************************ * * Rebuilding a case expression * * ************************************************************************ Note [Case elimination] ~~~~~~~~~~~~~~~~~~~~~~~ The case-elimination transformation discards redundant case expressions. Start with a simple situation: case x# of ===> let y# = x# in e y# -> e (when x#, y# are of primitive type, of course). We can't (in general) do this for algebraic cases, because we might turn bottom into non-bottom! The code in SimplUtils.prepareAlts has the effect of generalise this idea to look for a case where we're scrutinising a variable, and we know that only the default case can match. For example: case x of 0# -> ... DEFAULT -> ...(case x of 0# -> ... DEFAULT -> ...) ... Here the inner case is first trimmed to have only one alternative, the DEFAULT, after which it's an instance of the previous case. This really only shows up in eliminating error-checking code. Note that SimplUtils.mkCase combines identical RHSs. So case e of ===> case e of DEFAULT -> r True -> r False -> r Now again the case may be elminated by the CaseElim transformation. This includes things like (==# a# b#)::Bool so that we simplify case ==# a# b# of { True -> x; False -> x } to just x This particular example shows up in default methods for comparison operations (e.g. in (>=) for Int.Int32) Note [Case elimination: lifted case] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ If a case over a lifted type has a single alternative, and is being used as a strict 'let' (all isDeadBinder bndrs), we may want to do this transformation: case e of r ===> let r = e in ...r... _ -> ...r... (a) 'e' is already evaluated (it may so if e is a variable) Specifically we check (exprIsHNF e). In this case we can just allocate the WHNF directly with a let. or (b) 'x' is not used at all and e is ok-for-speculation The ok-for-spec bit checks that we don't lose any exceptions or divergence. NB: it'd be *sound* to switch from case to let if the scrutinee was not yet WHNF but was guaranteed to converge; but sticking with case means we won't build a thunk or (c) 'x' is used strictly in the body, and 'e' is a variable Then we can just substitute 'e' for 'x' in the body. See Note [Eliminating redundant seqs] For (b), the "not used at all" test is important. Consider case (case a ># b of { True -> (p,q); False -> (q,p) }) of r -> blah The scrutinee is ok-for-speculation (it looks inside cases), but we do not want to transform to let r = case a ># b of { True -> (p,q); False -> (q,p) } in blah because that builds an unnecessary thunk. Note [Eliminating redundant seqs] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ If we have this: case x of r { _ -> ..r.. } where 'r' is used strictly in (..r..), the case is effectively a 'seq' on 'x', but since 'r' is used strictly anyway, we can safely transform to (...x...) Note that this can change the error behaviour. For example, we might transform case x of { _ -> error "bad" } --> error "bad" which is might be puzzling if 'x' currently lambda-bound, but later gets let-bound to (error "good"). Nevertheless, the paper "A semantics for imprecise exceptions" allows this transformation. If you want to fix the evaluation order, use 'pseq'. See Trac #8900 for an example where the loss of this transformation bit us in practice. See also Note [Empty case alternatives] in CoreSyn. Just for reference, the original code (added Jan 13) looked like this: || case_bndr_evald_next rhs case_bndr_evald_next :: CoreExpr -> Bool -- See Note [Case binder next] case_bndr_evald_next (Var v) = v == case_bndr case_bndr_evald_next (Cast e _) = case_bndr_evald_next e case_bndr_evald_next (App e _) = case_bndr_evald_next e case_bndr_evald_next (Case e _ _ _) = case_bndr_evald_next e case_bndr_evald_next _ = False (This came up when fixing Trac #7542. See also Note [Eta reduction of an eval'd function] in CoreUtils.) Note [Case elimination: unlifted case] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Consider case a +# b of r -> ...r... Then we do case-elimination (to make a let) followed by inlining, to get .....(a +# b).... If we have case indexArray# a i of r -> ...r... we might like to do the same, and inline the (indexArray# a i). But indexArray# is not okForSpeculation, so we don't build a let in rebuildCase (lest it get floated *out*), so the inlining doesn't happen either. This really isn't a big deal I think. The let can be Further notes about case elimination ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Consider: test :: Integer -> IO () test = print Turns out that this compiles to: Print.test = \ eta :: Integer eta1 :: Void# -> case PrelNum.< eta PrelNum.zeroInteger of wild { __DEFAULT -> case hPutStr stdout (PrelNum.jtos eta ($w[] @ Char)) eta1 of wild1 { (# new_s, a4 #) -> PrelIO.lvl23 new_s }} Notice the strange '<' which has no effect at all. This is a funny one. It started like this: f x y = if x < 0 then jtos x else if y==0 then "" else jtos x At a particular call site we have (f v 1). So we inline to get if v < 0 then jtos x else if 1==0 then "" else jtos x Now simplify the 1==0 conditional: if v<0 then jtos v else jtos v Now common-up the two branches of the case: case (v<0) of DEFAULT -> jtos v Why don't we drop the case? Because it's strict in v. It's technically wrong to drop even unnecessary evaluations, and in practice they may be a result of 'seq' so we *definitely* don't want to drop those. I don't really know how to improve this situation. -} --------------------------------------------------------- -- Eliminate the case if possible rebuildCase, reallyRebuildCase :: SimplEnv -> OutExpr -- Scrutinee -> InId -- Case binder -> [InAlt] -- Alternatives (inceasing order) -> SimplCont -> SimplM (SimplEnv, OutExpr) -------------------------------------------------- -- 1. Eliminate the case if there's a known constructor -------------------------------------------------- rebuildCase env scrut case_bndr alts cont | Lit lit <- scrut -- No need for same treatment as constructors -- because literals are inlined more vigorously , not (litIsLifted lit) = do { tick (KnownBranch case_bndr) ; case findAlt (LitAlt lit) alts of Nothing -> missingAlt env case_bndr alts cont Just (_, bs, rhs) -> simple_rhs bs rhs } | Just (con, ty_args, other_args) <- exprIsConApp_maybe (getUnfoldingInRuleMatch env) scrut -- Works when the scrutinee is a variable with a known unfolding -- as well as when it's an explicit constructor application = do { tick (KnownBranch case_bndr) ; case findAlt (DataAlt con) alts of Nothing -> missingAlt env case_bndr alts cont Just (DEFAULT, bs, rhs) -> simple_rhs bs rhs Just (_, bs, rhs) -> knownCon env scrut con ty_args other_args case_bndr bs rhs cont } where simple_rhs bs rhs = ASSERT( null bs ) do { env' <- simplNonRecX env case_bndr scrut -- scrut is a constructor application, -- hence satisfies let/app invariant ; simplExprF env' rhs cont } -------------------------------------------------- -- 2. Eliminate the case if scrutinee is evaluated -------------------------------------------------- rebuildCase env scrut case_bndr alts@[(_, bndrs, rhs)] cont -- See if we can get rid of the case altogether -- See Note [Case elimination] -- mkCase made sure that if all the alternatives are equal, -- then there is now only one (DEFAULT) rhs -- 2a. Dropping the case altogether, if -- a) it binds nothing (so it's really just a 'seq') -- b) evaluating the scrutinee has no side effects | is_plain_seq , exprOkForSideEffects scrut -- The entire case is dead, so we can drop it -- if the scrutinee converges without having imperative -- side effects or raising a Haskell exception -- See Note [PrimOp can_fail and has_side_effects] in PrimOp = simplExprF env rhs cont -- 2b. Turn the case into a let, if -- a) it binds only the case-binder -- b) unlifted case: the scrutinee is ok-for-speculation -- lifted case: the scrutinee is in HNF (or will later be demanded) | all_dead_bndrs , if is_unlifted then exprOkForSpeculation scrut -- See Note [Case elimination: unlifted case] else exprIsHNF scrut -- See Note [Case elimination: lifted case] || scrut_is_demanded_var scrut = do { tick (CaseElim case_bndr) ; env' <- simplNonRecX env case_bndr scrut ; simplExprF env' rhs cont } -- 2c. Try the seq rules if -- a) it binds only the case binder -- b) a rule for seq applies -- See Note [User-defined RULES for seq] in MkId | is_plain_seq = do { let scrut_ty = exprType scrut rhs_ty = substTy env (exprType rhs) out_args = [ TyArg { as_arg_ty = scrut_ty , as_hole_ty = seq_id_ty } , TyArg { as_arg_ty = rhs_ty , as_hole_ty = applyTy seq_id_ty scrut_ty } , ValArg scrut] rule_cont = ApplyToVal { sc_dup = NoDup, sc_arg = rhs , sc_env = env, sc_cont = cont } env' = zapSubstEnv env -- Lazily evaluated, so we don't do most of this ; rule_base <- getSimplRules ; mb_rule <- tryRules env' (getRules rule_base seqId) seqId out_args rule_cont ; case mb_rule of Just (rule_rhs, cont') -> simplExprF env' rule_rhs cont' Nothing -> reallyRebuildCase env scrut case_bndr alts cont } where is_unlifted = isUnLiftedType (idType case_bndr) all_dead_bndrs = all isDeadBinder bndrs -- bndrs are [InId] is_plain_seq = all_dead_bndrs && isDeadBinder case_bndr -- Evaluation *only* for effect seq_id_ty = idType seqId scrut_is_demanded_var :: CoreExpr -> Bool -- See Note [Eliminating redundant seqs] scrut_is_demanded_var (Cast s _) = scrut_is_demanded_var s scrut_is_demanded_var (Var _) = isStrictDmd (idDemandInfo case_bndr) scrut_is_demanded_var _ = False rebuildCase env scrut case_bndr alts cont = reallyRebuildCase env scrut case_bndr alts cont -------------------------------------------------- -- 3. Catch-all case -------------------------------------------------- reallyRebuildCase env scrut case_bndr alts cont = do { -- Prepare the continuation; -- The new subst_env is in place (env', dup_cont, nodup_cont) <- prepareCaseCont env alts cont -- Simplify the alternatives ; (scrut', case_bndr', alts') <- simplAlts env' scrut case_bndr alts dup_cont ; dflags <- getDynFlags ; let alts_ty' = contResultType dup_cont ; case_expr <- mkCase dflags scrut' case_bndr' alts_ty' alts' -- Notice that rebuild gets the in-scope set from env', not alt_env -- (which in any case is only build in simplAlts) -- The case binder *not* scope over the whole returned case-expression ; rebuild env' case_expr nodup_cont } {- simplCaseBinder checks whether the scrutinee is a variable, v. If so, try to eliminate uses of v in the RHSs in favour of case_bndr; that way, there's a chance that v will now only be used once, and hence inlined. Historical note: we use to do the "case binder swap" in the Simplifier so there were additional complications if the scrutinee was a variable. Now the binder-swap stuff is done in the occurrence analyer; see OccurAnal Note [Binder swap]. Note [knownCon occ info] ~~~~~~~~~~~~~~~~~~~~~~~~ If the case binder is not dead, then neither are the pattern bound variables: case <any> of x { (a,b) -> case x of { (p,q) -> p } } Here (a,b) both look dead, but come alive after the inner case is eliminated. The point is that we bring into the envt a binding let x = (a,b) after the outer case, and that makes (a,b) alive. At least we do unless the case binder is guaranteed dead. Note [Case alternative occ info] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ When we are simply reconstructing a case (the common case), we always zap the occurrence info on the binders in the alternatives. Even if the case binder is dead, the scrutinee is usually a variable, and *that* can bring the case-alternative binders back to life. See Note [Add unfolding for scrutinee] Note [Improving seq] ~~~~~~~~~~~~~~~~~~~ Consider type family F :: * -> * type instance F Int = Int ... case e of x { DEFAULT -> rhs } ... where x::F Int. Then we'd like to rewrite (F Int) to Int, getting case e `cast` co of x'::Int I# x# -> let x = x' `cast` sym co in rhs so that 'rhs' can take advantage of the form of x'. Notice that Note [Case of cast] (in OccurAnal) may then apply to the result. Nota Bene: We only do the [Improving seq] transformation if the case binder 'x' is actually used in the rhs; that is, if the case is *not* a *pure* seq. a) There is no point in adding the cast to a pure seq. b) There is a good reason not to: doing so would interfere with seq rules (Note [Built-in RULES for seq] in MkId). In particular, this [Improving seq] thing *adds* a cast while [Built-in RULES for seq] *removes* one, so they just flip-flop. You might worry about case v of x { __DEFAULT -> ... case (v `cast` co) of y { I# -> ... }} This is a pure seq (since x is unused), so [Improving seq] won't happen. But it's ok: the simplifier will replace 'v' by 'x' in the rhs to get case v of x { __DEFAULT -> ... case (x `cast` co) of y { I# -> ... }} Now the outer case is not a pure seq, so [Improving seq] will happen, and then the inner case will disappear. The need for [Improving seq] showed up in Roman's experiments. Example: foo :: F Int -> Int -> Int foo t n = t `seq` bar n where bar 0 = 0 bar n = bar (n - case t of TI i -> i) Here we'd like to avoid repeated evaluating t inside the loop, by taking advantage of the `seq`. At one point I did transformation in LiberateCase, but it's more robust here. (Otherwise, there's a danger that we'll simply drop the 'seq' altogether, before LiberateCase gets to see it.) -} simplAlts :: SimplEnv -> OutExpr -> InId -- Case binder -> [InAlt] -- Non-empty -> SimplCont -> SimplM (OutExpr, OutId, [OutAlt]) -- Includes the continuation -- Like simplExpr, this just returns the simplified alternatives; -- it does not return an environment -- The returned alternatives can be empty, none are possible simplAlts env scrut case_bndr alts cont' = do { let env0 = zapFloats env ; (env1, case_bndr1) <- simplBinder env0 case_bndr ; fam_envs <- getFamEnvs ; (alt_env', scrut', case_bndr') <- improveSeq fam_envs env1 scrut case_bndr case_bndr1 alts ; (imposs_deflt_cons, in_alts) <- prepareAlts scrut' case_bndr' alts -- NB: it's possible that the returned in_alts is empty: this is handled -- by the caller (rebuildCase) in the missingAlt function ; alts' <- mapM (simplAlt alt_env' (Just scrut') imposs_deflt_cons case_bndr' cont') in_alts ; -- pprTrace "simplAlts" (ppr case_bndr $$ ppr alts_ty $$ ppr alts_ty' $$ ppr alts $$ ppr cont') $ return (scrut', case_bndr', alts') } ------------------------------------ improveSeq :: (FamInstEnv, FamInstEnv) -> SimplEnv -> OutExpr -> InId -> OutId -> [InAlt] -> SimplM (SimplEnv, OutExpr, OutId) -- Note [Improving seq] improveSeq fam_envs env scrut case_bndr case_bndr1 [(DEFAULT,_,_)] | not (isDeadBinder case_bndr) -- Not a pure seq! See Note [Improving seq] , Just (co, ty2) <- topNormaliseType_maybe fam_envs (idType case_bndr1) = do { case_bndr2 <- newId (fsLit "nt") ty2 ; let rhs = DoneEx (Var case_bndr2 `Cast` mkSymCo co) env2 = extendIdSubst env case_bndr rhs ; return (env2, scrut `Cast` co, case_bndr2) } improveSeq _ env scrut _ case_bndr1 _ = return (env, scrut, case_bndr1) ------------------------------------ simplAlt :: SimplEnv -> Maybe OutExpr -- The scrutinee -> [AltCon] -- These constructors can't be present when -- matching the DEFAULT alternative -> OutId -- The case binder -> SimplCont -> InAlt -> SimplM OutAlt simplAlt env _ imposs_deflt_cons case_bndr' cont' (DEFAULT, bndrs, rhs) = ASSERT( null bndrs ) do { let env' = addBinderUnfolding env case_bndr' (mkOtherCon imposs_deflt_cons) -- Record the constructors that the case-binder *can't* be. ; rhs' <- simplExprC env' rhs cont' ; return (DEFAULT, [], rhs') } simplAlt env scrut' _ case_bndr' cont' (LitAlt lit, bndrs, rhs) = ASSERT( null bndrs ) do { env' <- addAltUnfoldings env scrut' case_bndr' (Lit lit) ; rhs' <- simplExprC env' rhs cont' ; return (LitAlt lit, [], rhs') } simplAlt env scrut' _ case_bndr' cont' (DataAlt con, vs, rhs) = do { -- Deal with the pattern-bound variables -- Mark the ones that are in ! positions in the -- data constructor as certainly-evaluated. -- NB: simplLamBinders preserves this eval info ; let vs_with_evals = add_evals (dataConRepStrictness con) ; (env', vs') <- simplLamBndrs env vs_with_evals -- Bind the case-binder to (con args) ; let inst_tys' = tyConAppArgs (idType case_bndr') con_app :: OutExpr con_app = mkConApp2 con inst_tys' vs' ; env'' <- addAltUnfoldings env' scrut' case_bndr' con_app ; rhs' <- simplExprC env'' rhs cont' ; return (DataAlt con, vs', rhs') } where -- add_evals records the evaluated-ness of the bound variables of -- a case pattern. This is *important*. Consider -- data T = T !Int !Int -- -- case x of { T a b -> T (a+1) b } -- -- We really must record that b is already evaluated so that we don't -- go and re-evaluate it when constructing the result. -- See Note [Data-con worker strictness] in MkId.hs add_evals the_strs = go vs the_strs where go [] [] = [] go (v:vs') strs | isTyVar v = v : go vs' strs go (v:vs') (str:strs) | isMarkedStrict str = evald_v : go vs' strs | otherwise = zapped_v : go vs' strs where zapped_v = zapIdOccInfo v -- See Note [Case alternative occ info] evald_v = zapped_v `setIdUnfolding` evaldUnfolding go _ _ = pprPanic "cat_evals" (ppr con $$ ppr vs $$ ppr the_strs) addAltUnfoldings :: SimplEnv -> Maybe OutExpr -> OutId -> OutExpr -> SimplM SimplEnv addAltUnfoldings env scrut case_bndr con_app = do { dflags <- getDynFlags ; let con_app_unf = mkSimpleUnfolding dflags con_app env1 = addBinderUnfolding env case_bndr con_app_unf -- See Note [Add unfolding for scrutinee] env2 = case scrut of Just (Var v) -> addBinderUnfolding env1 v con_app_unf Just (Cast (Var v) co) -> addBinderUnfolding env1 v $ mkSimpleUnfolding dflags (Cast con_app (mkSymCo co)) _ -> env1 ; traceSmpl "addAltUnf" (vcat [ppr case_bndr <+> ppr scrut, ppr con_app]) ; return env2 } addBinderUnfolding :: SimplEnv -> Id -> Unfolding -> SimplEnv addBinderUnfolding env bndr unf | debugIsOn, Just tmpl <- maybeUnfoldingTemplate unf = WARN( not (eqType (idType bndr) (exprType tmpl)), ppr bndr $$ ppr (idType bndr) $$ ppr tmpl $$ ppr (exprType tmpl) ) modifyInScope env (bndr `setIdUnfolding` unf) | otherwise = modifyInScope env (bndr `setIdUnfolding` unf) zapBndrOccInfo :: Bool -> Id -> Id -- Consider case e of b { (a,b) -> ... } -- Then if we bind b to (a,b) in "...", and b is not dead, -- then we must zap the deadness info on a,b zapBndrOccInfo keep_occ_info pat_id | keep_occ_info = pat_id | otherwise = zapIdOccInfo pat_id {- Note [Add unfolding for scrutinee] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In general it's unlikely that a variable scrutinee will appear in the case alternatives case x of { ...x unlikely to appear... } because the binder-swap in OccAnal has got rid of all such occcurrences See Note [Binder swap] in OccAnal. BUT it is still VERY IMPORTANT to add a suitable unfolding for a variable scrutinee, in simplAlt. Here's why case x of y (a,b) -> case b of c I# v -> ...(f y)... There is no occurrence of 'b' in the (...(f y)...). But y gets the unfolding (a,b), and *that* mentions b. If f has a RULE RULE f (p, I# q) = ... we want that rule to match, so we must extend the in-scope env with a suitable unfolding for 'y'. It's *essential* for rule matching; but it's also good for case-elimintation -- suppose that 'f' was inlined and did multi-level case analysis, then we'd solve it in one simplifier sweep instead of two. Exactly the same issue arises in SpecConstr; see Note [Add scrutinee to ValueEnv too] in SpecConstr HOWEVER, given case x of y { Just a -> r1; Nothing -> r2 } we do not want to add the unfolding x -> y to 'x', which might seem cool, since 'y' itself has different unfoldings in r1 and r2. Reason: if we did that, we'd have to zap y's deadness info and that is a very useful piece of information. So instead we add the unfolding x -> Just a, and x -> Nothing in the respective RHSs. ************************************************************************ * * \subsection{Known constructor} * * ************************************************************************ We are a bit careful with occurrence info. Here's an example (\x* -> case x of (a*, b) -> f a) (h v, e) where the * means "occurs once". This effectively becomes case (h v, e) of (a*, b) -> f a) and then let a* = h v; b = e in f a and then f (h v) All this should happen in one sweep. -} knownCon :: SimplEnv -> OutExpr -- The scrutinee -> DataCon -> [OutType] -> [OutExpr] -- The scrutinee (in pieces) -> InId -> [InBndr] -> InExpr -- The alternative -> SimplCont -> SimplM (SimplEnv, OutExpr) knownCon env scrut dc dc_ty_args dc_args bndr bs rhs cont = do { env' <- bind_args env bs dc_args ; env'' <- bind_case_bndr env' ; simplExprF env'' rhs cont } where zap_occ = zapBndrOccInfo (isDeadBinder bndr) -- bndr is an InId -- Ugh! bind_args env' [] _ = return env' bind_args env' (b:bs') (Type ty : args) = ASSERT( isTyVar b ) bind_args (extendTvSubst env' b ty) bs' args bind_args env' (b:bs') (arg : args) = ASSERT( isId b ) do { let b' = zap_occ b -- Note that the binder might be "dead", because it doesn't -- occur in the RHS; and simplNonRecX may therefore discard -- it via postInlineUnconditionally. -- Nevertheless we must keep it if the case-binder is alive, -- because it may be used in the con_app. See Note [knownCon occ info] ; env'' <- simplNonRecX env' b' arg -- arg satisfies let/app invariant ; bind_args env'' bs' args } bind_args _ _ _ = pprPanic "bind_args" $ ppr dc $$ ppr bs $$ ppr dc_args $$ text "scrut:" <+> ppr scrut -- It's useful to bind bndr to scrut, rather than to a fresh -- binding x = Con arg1 .. argn -- because very often the scrut is a variable, so we avoid -- creating, and then subsequently eliminating, a let-binding -- BUT, if scrut is a not a variable, we must be careful -- about duplicating the arg redexes; in that case, make -- a new con-app from the args bind_case_bndr env | isDeadBinder bndr = return env | exprIsTrivial scrut = return (extendIdSubst env bndr (DoneEx scrut)) | otherwise = do { dc_args <- mapM (simplVar env) bs -- dc_ty_args are aready OutTypes, -- but bs are InBndrs ; let con_app = Var (dataConWorkId dc) `mkTyApps` dc_ty_args `mkApps` dc_args ; simplNonRecX env bndr con_app } ------------------- missingAlt :: SimplEnv -> Id -> [InAlt] -> SimplCont -> SimplM (SimplEnv, OutExpr) -- This isn't strictly an error, although it is unusual. -- It's possible that the simplifer might "see" that -- an inner case has no accessible alternatives before -- it "sees" that the entire branch of an outer case is -- inaccessible. So we simply put an error case here instead. missingAlt env case_bndr _ cont = WARN( True, ptext (sLit "missingAlt") <+> ppr case_bndr ) return (env, mkImpossibleExpr (contResultType cont)) {- ************************************************************************ * * \subsection{Duplicating continuations} * * ************************************************************************ -} prepareCaseCont :: SimplEnv -> [InAlt] -> SimplCont -> SimplM (SimplEnv, SimplCont, -- Dupable part SimplCont) -- Non-dupable part -- We are considering -- K[case _ of { p1 -> r1; ...; pn -> rn }] -- where K is some enclosing continuation for the case -- Goal: split K into two pieces Kdup,Knodup so that -- a) Kdup can be duplicated -- b) Knodup[Kdup[e]] = K[e] -- The idea is that we'll transform thus: -- Knodup[ (case _ of { p1 -> Kdup[r1]; ...; pn -> Kdup[rn] } -- -- We may also return some extra bindings in SimplEnv (that scope over -- the entire continuation) -- -- When case-of-case is off, just make the entire continuation non-dupable prepareCaseCont env alts cont | not (sm_case_case (getMode env)) = return (env, mkBoringStop (contHoleType cont), cont) | not (many_alts alts) = return (env, cont, mkBoringStop (contResultType cont)) | otherwise = mkDupableCont env cont where many_alts :: [InAlt] -> Bool -- True iff strictly > 1 non-bottom alternative many_alts [] = False -- See Note [Bottom alternatives] many_alts [_] = False many_alts (alt:alts) | is_bot_alt alt = many_alts alts | otherwise = not (all is_bot_alt alts) is_bot_alt (_,_,rhs) = exprIsBottom rhs {- Note [Bottom alternatives] ~~~~~~~~~~~~~~~~~~~~~~~~~~ When we have case (case x of { A -> error .. ; B -> e; C -> error ..) of alts then we can just duplicate those alts because the A and C cases will disappear immediately. This is more direct than creating join points and inlining them away; and in some cases we would not even create the join points (see Note [Single-alternative case]) and we would keep the case-of-case which is silly. See Trac #4930. -} mkDupableCont :: SimplEnv -> SimplCont -> SimplM (SimplEnv, SimplCont, SimplCont) mkDupableCont env cont | contIsDupable cont = return (env, cont, mkBoringStop (contResultType cont)) mkDupableCont _ (Stop {}) = panic "mkDupableCont" -- Handled by previous eqn mkDupableCont env (CastIt ty cont) = do { (env', dup, nodup) <- mkDupableCont env cont ; return (env', CastIt ty dup, nodup) } -- Duplicating ticks for now, not sure if this is good or not mkDupableCont env cont@(TickIt{}) = return (env, mkBoringStop (contHoleType cont), cont) mkDupableCont env cont@(StrictBind {}) = return (env, mkBoringStop (contHoleType cont), cont) -- See Note [Duplicating StrictBind] mkDupableCont env (StrictArg info cci cont) -- See Note [Duplicating StrictArg] = do { (env', dup, nodup) <- mkDupableCont env cont ; (env'', args') <- mapAccumLM makeTrivialArg env' (ai_args info) ; return (env'', StrictArg (info { ai_args = args' }) cci dup, nodup) } mkDupableCont env cont@(ApplyToTy { sc_cont = tail }) = do { (env', dup_cont, nodup_cont) <- mkDupableCont env tail ; return (env', cont { sc_cont = dup_cont }, nodup_cont ) } mkDupableCont env (ApplyToVal { sc_arg = arg, sc_dup = dup, sc_env = se, sc_cont = cont }) = -- e.g. [...hole...] (...arg...) -- ==> -- let a = ...arg... -- in [...hole...] a do { (env', dup_cont, nodup_cont) <- mkDupableCont env cont ; (_, se', arg') <- simplArg env' dup se arg ; (env'', arg'') <- makeTrivial NotTopLevel env' arg' ; let app_cont = ApplyToVal { sc_arg = arg'', sc_env = se' , sc_dup = OkToDup, sc_cont = dup_cont } ; return (env'', app_cont, nodup_cont) } mkDupableCont env cont@(Select { sc_bndr = case_bndr, sc_alts = [(_, bs, _rhs)] }) -- See Note [Single-alternative case] -- | not (exprIsDupable rhs && contIsDupable case_cont) -- | not (isDeadBinder case_bndr) | all isDeadBinder bs -- InIds && not (isUnLiftedType (idType case_bndr)) -- Note [Single-alternative-unlifted] = return (env, mkBoringStop (contHoleType cont), cont) mkDupableCont env (Select { sc_bndr = case_bndr, sc_alts = alts , sc_env = se, sc_cont = cont }) = -- e.g. (case [...hole...] of { pi -> ei }) -- ===> -- let ji = \xij -> ei -- in case [...hole...] of { pi -> ji xij } do { tick (CaseOfCase case_bndr) ; (env', dup_cont, nodup_cont) <- prepareCaseCont env alts cont -- NB: We call prepareCaseCont here. If there is only one -- alternative, then dup_cont may be big, but that's ok -- because we push it into the single alternative, and then -- use mkDupableAlt to turn that simplified alternative into -- a join point if it's too big to duplicate. -- And this is important: see Note [Fusing case continuations] ; let alt_env = se `setInScope` env' ; (alt_env', case_bndr') <- simplBinder alt_env case_bndr ; alts' <- mapM (simplAlt alt_env' Nothing [] case_bndr' dup_cont) alts -- Safe to say that there are no handled-cons for the DEFAULT case -- NB: simplBinder does not zap deadness occ-info, so -- a dead case_bndr' will still advertise its deadness -- This is really important because in -- case e of b { (# p,q #) -> ... } -- b is always dead, and indeed we are not allowed to bind b to (# p,q #), -- which might happen if e was an explicit unboxed pair and b wasn't marked dead. -- In the new alts we build, we have the new case binder, so it must retain -- its deadness. -- NB: we don't use alt_env further; it has the substEnv for -- the alternatives, and we don't want that ; (env'', alts'') <- mkDupableAlts env' case_bndr' alts' ; return (env'', -- Note [Duplicated env] Select { sc_dup = OkToDup , sc_bndr = case_bndr', sc_alts = alts'' , sc_env = zapSubstEnv env'' , sc_cont = mkBoringStop (contHoleType nodup_cont) }, nodup_cont) } mkDupableAlts :: SimplEnv -> OutId -> [InAlt] -> SimplM (SimplEnv, [InAlt]) -- Absorbs the continuation into the new alternatives mkDupableAlts env case_bndr' the_alts = go env the_alts where go env0 [] = return (env0, []) go env0 (alt:alts) = do { (env1, alt') <- mkDupableAlt env0 case_bndr' alt ; (env2, alts') <- go env1 alts ; return (env2, alt' : alts' ) } mkDupableAlt :: SimplEnv -> OutId -> (AltCon, [CoreBndr], CoreExpr) -> SimplM (SimplEnv, (AltCon, [CoreBndr], CoreExpr)) mkDupableAlt env case_bndr (con, bndrs', rhs') = do dflags <- getDynFlags if exprIsDupable dflags rhs' -- Note [Small alternative rhs] then return (env, (con, bndrs', rhs')) else do { let rhs_ty' = exprType rhs' scrut_ty = idType case_bndr case_bndr_w_unf = case con of DEFAULT -> case_bndr DataAlt dc -> setIdUnfolding case_bndr unf where -- See Note [Case binders and join points] unf = mkInlineUnfolding Nothing rhs rhs = mkConApp2 dc (tyConAppArgs scrut_ty) bndrs' LitAlt {} -> WARN( True, ptext (sLit "mkDupableAlt") <+> ppr case_bndr <+> ppr con ) case_bndr -- The case binder is alive but trivial, so why has -- it not been substituted away? used_bndrs' | isDeadBinder case_bndr = filter abstract_over bndrs' | otherwise = bndrs' ++ [case_bndr_w_unf] abstract_over bndr | isTyVar bndr = True -- Abstract over all type variables just in case | otherwise = not (isDeadBinder bndr) -- The deadness info on the new Ids is preserved by simplBinders ; (final_bndrs', final_args) -- Note [Join point abstraction] <- if (any isId used_bndrs') then return (used_bndrs', varsToCoreExprs used_bndrs') else do { rw_id <- newId (fsLit "w") voidPrimTy ; return ([setOneShotLambda rw_id], [Var voidPrimId]) } ; join_bndr <- newId (fsLit "$j") (mkPiTypes final_bndrs' rhs_ty') -- Note [Funky mkPiTypes] ; let -- We make the lambdas into one-shot-lambdas. The -- join point is sure to be applied at most once, and doing so -- prevents the body of the join point being floated out by -- the full laziness pass really_final_bndrs = map one_shot final_bndrs' one_shot v | isId v = setOneShotLambda v | otherwise = v join_rhs = mkLams really_final_bndrs rhs' join_arity = exprArity join_rhs join_call = mkApps (Var join_bndr) final_args ; env' <- addPolyBind NotTopLevel env (NonRec (join_bndr `setIdArity` join_arity) join_rhs) ; return (env', (con, bndrs', join_call)) } -- See Note [Duplicated env] {- Note [Fusing case continuations] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ It's important to fuse two successive case continuations when the first has one alternative. That's why we call prepareCaseCont here. Consider this, which arises from thunk splitting (see Note [Thunk splitting] in WorkWrap): let x* = case (case v of {pn -> rn}) of I# a -> I# a in body The simplifier will find (Var v) with continuation Select (pn -> rn) ( Select [I# a -> I# a] ( StrictBind body Stop So we'll call mkDupableCont on Select [I# a -> I# a] (StrictBind body Stop) There is just one alternative in the first Select, so we want to simplify the rhs (I# a) with continuation (StricgtBind body Stop) Supposing that body is big, we end up with let $j a = <let x = I# a in body> in case v of { pn -> case rn of I# a -> $j a } This is just what we want because the rn produces a box that the case rn cancels with. See Trac #4957 a fuller example. Note [Case binders and join points] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Consider this case (case .. ) of c { I# c# -> ....c.... If we make a join point with c but not c# we get $j = \c -> ....c.... But if later inlining scrutines the c, thus $j = \c -> ... case c of { I# y -> ... } ... we won't see that 'c' has already been scrutinised. This actually happens in the 'tabulate' function in wave4main, and makes a significant difference to allocation. An alternative plan is this: $j = \c# -> let c = I# c# in ...c.... but that is bad if 'c' is *not* later scrutinised. So instead we do both: we pass 'c' and 'c#' , and record in c's inlining (a stable unfolding) that it's really I# c#, thus $j = \c# -> \c[=I# c#] -> ...c.... Absence analysis may later discard 'c'. NB: take great care when doing strictness analysis; see Note [Lamba-bound unfoldings] in DmdAnal. Also note that we can still end up passing stuff that isn't used. Before strictness analysis we have let $j x y c{=(x,y)} = (h c, ...) in ... After strictness analysis we see that h is strict, we end up with let $j x y c{=(x,y)} = ($wh x y, ...) and c is unused. Note [Duplicated env] ~~~~~~~~~~~~~~~~~~~~~ Some of the alternatives are simplified, but have not been turned into a join point So they *must* have an zapped subst-env. So we can't use completeNonRecX to bind the join point, because it might to do PostInlineUnconditionally, and we'd lose that when zapping the subst-env. We could have a per-alt subst-env, but zapping it (as we do in mkDupableCont, the Select case) is safe, and at worst delays the join-point inlining. Note [Small alternative rhs] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~ It is worth checking for a small RHS because otherwise we get extra let bindings that may cause an extra iteration of the simplifier to inline back in place. Quite often the rhs is just a variable or constructor. The Ord instance of Maybe in PrelMaybe.hs, for example, took several extra iterations because the version with the let bindings looked big, and so wasn't inlined, but after the join points had been inlined it looked smaller, and so was inlined. NB: we have to check the size of rhs', not rhs. Duplicating a small InAlt might invalidate occurrence information However, if it *is* dupable, we return the *un* simplified alternative, because otherwise we'd need to pair it up with an empty subst-env.... but we only have one env shared between all the alts. (Remember we must zap the subst-env before re-simplifying something). Rather than do this we simply agree to re-simplify the original (small) thing later. Note [Funky mkPiTypes] ~~~~~~~~~~~~~~~~~~~~~~ Notice the funky mkPiTypes. If the contructor has existentials it's possible that the join point will be abstracted over type variables as well as term variables. Example: Suppose we have data T = forall t. C [t] Then faced with case (case e of ...) of C t xs::[t] -> rhs We get the join point let j :: forall t. [t] -> ... j = /\t \xs::[t] -> rhs in case (case e of ...) of C t xs::[t] -> j t xs Note [Join point abstraction] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Join points always have at least one value argument, for several reasons * If we try to lift a primitive-typed something out for let-binding-purposes, we will *caseify* it (!), with potentially-disastrous strictness results. So instead we turn it into a function: \v -> e where v::Void#. The value passed to this function is void, which generates (almost) no code. * CPR. We used to say "&& isUnLiftedType rhs_ty'" here, but now we make the join point into a function whenever used_bndrs' is empty. This makes the join-point more CPR friendly. Consider: let j = if .. then I# 3 else I# 4 in case .. of { A -> j; B -> j; C -> ... } Now CPR doesn't w/w j because it's a thunk, so that means that the enclosing function can't w/w either, which is a lose. Here's the example that happened in practice: kgmod :: Int -> Int -> Int kgmod x y = if x > 0 && y < 0 || x < 0 && y > 0 then 78 else 5 * Let-no-escape. We want a join point to turn into a let-no-escape so that it is implemented as a jump, and one of the conditions for LNE is that it's not updatable. In CoreToStg, see Note [What is a non-escaping let] * Floating. Since a join point will be entered once, no sharing is gained by floating out, but something might be lost by doing so because it might be allocated. I have seen a case alternative like this: True -> \v -> ... It's a bit silly to add the realWorld dummy arg in this case, making $j = \s v -> ... True -> $j s (the \v alone is enough to make CPR happy) but I think it's rare There's a slight infelicity here: we pass the overall case_bndr to all the join points if it's used in *any* RHS, because we don't know its usage in each RHS separately Note [Duplicating StrictArg] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~ The original plan had (where E is a big argument) e.g. f E [..hole..] ==> let $j = \a -> f E a in $j [..hole..] But this is terrible! Here's an example: && E (case x of { T -> F; F -> T }) Now, && is strict so we end up simplifying the case with an ArgOf continuation. If we let-bind it, we get let $j = \v -> && E v in simplExpr (case x of { T -> F; F -> T }) (ArgOf (\r -> $j r) And after simplifying more we get let $j = \v -> && E v in case x of { T -> $j F; F -> $j T } Which is a Very Bad Thing What we do now is this f E [..hole..] ==> let a = E in f a [..hole..] Now if the thing in the hole is a case expression (which is when we'll call mkDupableCont), we'll push the function call into the branches, which is what we want. Now RULES for f may fire, and call-pattern specialisation. Here's an example from Trac #3116 go (n+1) (case l of 1 -> bs' _ -> Chunk p fpc (o+1) (l-1) bs') If we can push the call for 'go' inside the case, we get call-pattern specialisation for 'go', which is *crucial* for this program. Here is the (&&) example: && E (case x of { T -> F; F -> T }) ==> let a = E in case x of { T -> && a F; F -> && a T } Much better! Notice that * Arguments to f *after* the strict one are handled by the ApplyToVal case of mkDupableCont. Eg f [..hole..] E * We can only do the let-binding of E because the function part of a StrictArg continuation is an explicit syntax tree. In earlier versions we represented it as a function (CoreExpr -> CoreEpxr) which we couldn't take apart. Do *not* duplicate StrictBind and StritArg continuations. We gain nothing by propagating them into the expressions, and we do lose a lot. The desire not to duplicate is the entire reason that mkDupableCont returns a pair of continuations. Note [Duplicating StrictBind] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Unlike StrictArg, there doesn't seem anything to gain from duplicating a StrictBind continuation, so we don't. Note [Single-alternative cases] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ This case is just like the ArgOf case. Here's an example: data T a = MkT !a ...(MkT (abs x))... Then we get case (case x of I# x' -> case x' <# 0# of True -> I# (negate# x') False -> I# x') of y { DEFAULT -> MkT y Because the (case x) has only one alternative, we'll transform to case x of I# x' -> case (case x' <# 0# of True -> I# (negate# x') False -> I# x') of y { DEFAULT -> MkT y But now we do *NOT* want to make a join point etc, giving case x of I# x' -> let $j = \y -> MkT y in case x' <# 0# of True -> $j (I# (negate# x')) False -> $j (I# x') In this case the $j will inline again, but suppose there was a big strict computation enclosing the orginal call to MkT. Then, it won't "see" the MkT any more, because it's big and won't get duplicated. And, what is worse, nothing was gained by the case-of-case transform. So, in circumstances like these, we don't want to build join points and push the outer case into the branches of the inner one. Instead, don't duplicate the continuation. When should we use this strategy? We should not use it on *every* single-alternative case: e.g. case (case ....) of (a,b) -> (# a,b #) Here we must push the outer case into the inner one! Other choices: * Match [(DEFAULT,_,_)], but in the common case of Int, the alternative-filling-in code turned the outer case into case (...) of y { I# _ -> MkT y } * Match on single alternative plus (not (isDeadBinder case_bndr)) Rationale: pushing the case inwards won't eliminate the construction. But there's a risk of case (...) of y { (a,b) -> let z=(a,b) in ... } Now y looks dead, but it'll come alive again. Still, this seems like the best option at the moment. * Match on single alternative plus (all (isDeadBinder bndrs)) Rationale: this is essentially seq. * Match when the rhs is *not* duplicable, and hence would lead to a join point. This catches the disaster-case above. We can test the *un-simplified* rhs, which is fine. It might get bigger or smaller after simplification; if it gets smaller, this case might fire next time round. NB also that we must test contIsDupable case_cont *too, because case_cont might be big! HOWEVER: I found that this version doesn't work well, because we can get let x = case (...) of { small } in ...case x... When x is inlined into its full context, we find that it was a bad idea to have pushed the outer case inside the (...) case. There is a cost to not doing case-of-case; see Trac #10626. Note [Single-alternative-unlifted] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Here's another single-alternative where we really want to do case-of-case: data Mk1 = Mk1 Int# | Mk2 Int# M1.f = \r [x_s74 y_s6X] case case y_s6X of tpl_s7m { M1.Mk1 ipv_s70 -> ipv_s70; M1.Mk2 ipv_s72 -> ipv_s72; } of wild_s7c { __DEFAULT -> case case x_s74 of tpl_s7n { M1.Mk1 ipv_s77 -> ipv_s77; M1.Mk2 ipv_s79 -> ipv_s79; } of wild1_s7b { __DEFAULT -> ==# [wild1_s7b wild_s7c]; }; }; So the outer case is doing *nothing at all*, other than serving as a join-point. In this case we really want to do case-of-case and decide whether to use a real join point or just duplicate the continuation: let $j s7c = case x of Mk1 ipv77 -> (==) s7c ipv77 Mk1 ipv79 -> (==) s7c ipv79 in case y of Mk1 ipv70 -> $j ipv70 Mk2 ipv72 -> $j ipv72 Hence: check whether the case binder's type is unlifted, because then the outer case is *not* a seq. ************************************************************************ * * Unfoldings * * ************************************************************************ -} simplLetUnfolding :: SimplEnv-> TopLevelFlag -> InId -> OutExpr -> Unfolding -> SimplM Unfolding simplLetUnfolding env top_lvl id new_rhs unf | isStableUnfolding unf = simplUnfolding env top_lvl id unf | otherwise = bottoming `seq` -- See Note [Force bottoming field] do { dflags <- getDynFlags ; return (mkUnfolding dflags InlineRhs (isTopLevel top_lvl) bottoming new_rhs) } -- We make an unfolding *even for loop-breakers*. -- Reason: (a) It might be useful to know that they are WHNF -- (b) In TidyPgm we currently assume that, if we want to -- expose the unfolding then indeed we *have* an unfolding -- to expose. (We could instead use the RHS, but currently -- we don't.) The simple thing is always to have one. where bottoming = isBottomingId id simplUnfolding :: SimplEnv-> TopLevelFlag -> InId -> Unfolding -> SimplM Unfolding -- Note [Setting the new unfolding] simplUnfolding env top_lvl id unf = case unf of NoUnfolding -> return unf OtherCon {} -> return unf DFunUnfolding { df_bndrs = bndrs, df_con = con, df_args = args } -> do { (env', bndrs') <- simplBinders rule_env bndrs ; args' <- mapM (simplExpr env') args ; return (mkDFunUnfolding bndrs' con args') } CoreUnfolding { uf_tmpl = expr, uf_src = src, uf_guidance = guide } | isStableSource src -> do { expr' <- simplExpr rule_env expr ; case guide of UnfWhen { ug_arity = arity, ug_unsat_ok = sat_ok } -- Happens for INLINE things -> let guide' = UnfWhen { ug_arity = arity, ug_unsat_ok = sat_ok , ug_boring_ok = inlineBoringOk expr' } -- Refresh the boring-ok flag, in case expr' -- has got small. This happens, notably in the inlinings -- for dfuns for single-method classes; see -- Note [Single-method classes] in TcInstDcls. -- A test case is Trac #4138 in return (mkCoreUnfolding src is_top_lvl expr' guide') -- See Note [Top-level flag on inline rules] in CoreUnfold _other -- Happens for INLINABLE things -> bottoming `seq` -- See Note [Force bottoming field] do { dflags <- getDynFlags ; return (mkUnfolding dflags src is_top_lvl bottoming expr') } } -- If the guidance is UnfIfGoodArgs, this is an INLINABLE -- unfolding, and we need to make sure the guidance is kept up -- to date with respect to any changes in the unfolding. | otherwise -> return noUnfolding -- Discard unstable unfoldings where bottoming = isBottomingId id is_top_lvl = isTopLevel top_lvl act = idInlineActivation id rule_env = updMode (updModeForStableUnfoldings act) env -- See Note [Simplifying inside stable unfoldings] in SimplUtils {- Note [Force bottoming field] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~ We need to force bottoming, or the new unfolding holds on to the old unfolding (which is part of the id). Note [Setting the new unfolding] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ * If there's an INLINE pragma, we simplify the RHS gently. Maybe we should do nothing at all, but simplifying gently might get rid of more crap. * If not, we make an unfolding from the new RHS. But *only* for non-loop-breakers. Making loop breakers not have an unfolding at all means that we can avoid tests in exprIsConApp, for example. This is important: if exprIsConApp says 'yes' for a recursive thing, then we can get into an infinite loop If there's an stable unfolding on a loop breaker (which happens for INLINEABLE), we hang on to the inlining. It's pretty dodgy, but the user did say 'INLINE'. May need to revisit this choice. ************************************************************************ * * Rules * * ************************************************************************ Note [Rules in a letrec] ~~~~~~~~~~~~~~~~~~~~~~~~ After creating fresh binders for the binders of a letrec, we substitute the RULES and add them back onto the binders; this is done *before* processing any of the RHSs. This is important. Manuel found cases where he really, really wanted a RULE for a recursive function to apply in that function's own right-hand side. See Note [Loop breaking and RULES] in OccAnal. -} addBndrRules :: SimplEnv -> InBndr -> OutBndr -> SimplM (SimplEnv, OutBndr) -- Rules are added back into the bin addBndrRules env in_id out_id | null old_rules = return (env, out_id) | otherwise = do { new_rules <- simplRules env (Just (idName out_id)) old_rules ; let final_id = out_id `setIdSpecialisation` mkSpecInfo new_rules ; return (modifyInScope env final_id, final_id) } where old_rules = specInfoRules (idSpecialisation in_id) simplRules :: SimplEnv -> Maybe Name -> [CoreRule] -> SimplM [CoreRule] simplRules env mb_new_nm rules = mapM simpl_rule rules where simpl_rule rule@(BuiltinRule {}) = return rule simpl_rule rule@(Rule { ru_bndrs = bndrs, ru_args = args , ru_fn = fn_name, ru_rhs = rhs , ru_act = act }) = do { (env, bndrs') <- simplBinders env bndrs ; let lhs_env = updMode updModeForRuleLHS env rhs_env = updMode (updModeForStableUnfoldings act) env ; args' <- mapM (simplExpr lhs_env) args ; rhs' <- simplExpr rhs_env rhs ; return (rule { ru_bndrs = bndrs' , ru_fn = mb_new_nm `orElse` fn_name , ru_args = args' , ru_rhs = rhs' }) }
acowley/ghc
compiler/simplCore/Simplify.hs
bsd-3-clause
123,147
20
25
37,712
15,225
8,071
7,154
-1
-1
{-# LANGUAGE TemplateHaskell, TypeFamilies, PolyKinds #-} module T8884 where import Language.Haskell.TH import System.IO type family Foo a = r | r -> a where Foo x = x type family Baz (a :: k) = (r :: k) | r -> a type instance Baz x = x $( do FamilyI foo@(ClosedTypeFamilyD _ tvbs1 res1 m_kind1 eqns1) [] <- reify ''Foo FamilyI baz@(OpenTypeFamilyD _ tvbs2 res2 m_kind2) [inst@(TySynInstD _ eqn2)] <- reify ''Baz runIO $ putStrLn $ pprint foo runIO $ putStrLn $ pprint baz runIO $ putStrLn $ pprint inst runIO $ hFlush stdout return [ ClosedTypeFamilyD (mkName "Foo'") tvbs1 res1 m_kind1 eqns1 , OpenTypeFamilyD (mkName "Baz'") tvbs2 res2 m_kind2 , TySynInstD (mkName "Baz'") eqn2 ] )
ghc-android/ghc
testsuite/tests/th/T8884.hs
bsd-3-clause
784
0
14
217
274
143
131
19
0
module Shaker.CommonUtilTest where import Shaker.CommonUtil import Control.Arrow import Data.List prop_separateEqualShouldNotSplitNubList :: [Int] -> Bool prop_separateEqualShouldNotSplitNubList lst | null lst = separateEqual lst == [] | otherwise = (nub >>> separateEqual >>> head) lst == nub lst prop_separateEqualShouldSplitListWithEqualTerms :: [Int] -> Bool prop_separateEqualShouldSplitListWithEqualTerms lst | null lst = True | otherwise = all isContainingUnique (separateEqual lst) isContainingUnique :: Eq a => [a] -> Bool isContainingUnique lst = length lst == (nub >>> length) lst
bonnefoa/Shaker
testsuite/tests/Shaker/CommonUtilTest.hs
isc
696
0
10
177
182
90
92
12
1
{-# LANGUAGE OverloadedStrings #-} {-# LANGUAGE GADTs #-} {-# LANGUAGE DataKinds #-} {-# LANGUAGE FlexibleContexts #-} {-# LANGUAGE TypeOperators #-} {-# LANGUAGE ScopedTypeVariables #-} module Controllers.Sessions where import Lucid import Lucid.Html5 import Control.Monad import Data.HVect import Web.Spock.Safe import Models import Config import Views.Sessions as Views new :: (ContainsGuest n xs, NotAuthenticated xs) => Action (HVect xs) a new = render $ Views.new [] create :: (ContainsGuest n xs, NotAuthenticated xs) => Action (HVect xs) a create = do uName <- param' "email" uPass <- param' "password" loginRes <- runSQL $ loginUser uName uPass case loginRes of Just userId -> do sid <- runSQL $ createSession userId writeSession (Just sid) redirect "/" Nothing -> do render $ Views.new ["Wrong username or password"] destroy :: Authenticated n xs => Action (HVect xs) a destroy = do (userId, _ :: User) <- liftM findFirst getContext runSQL $ killSessions userId writeSession Nothing redirect "/"
folsen/haskell-startapp
src/Controllers/Sessions.hs
mit
1,070
0
14
214
315
158
157
37
2
module Demos ( mainDemo ) where import Draw import Centres import Shape import Constructions -------------------------------------------------------------------------------- type Demo = Triangle -> IO () mainDemo :: Demo mainDemo = demoNapoleonTriangle True -------------------------------------------------------------------------------- demoCentroid :: Demo demoCentroid t@(Triangle a b c) = let c' = centroid t in do draw white t draw red $ t `ceviansThrough` c' draw red c' demoSymmedian :: Demo demoSymmedian t@(Triangle a b c) = let c' = symmedianpoint t in do draw white t draw red $ t `ceviansThrough` c' draw red c' demoOrthocentre :: Demo demoOrthocentre t@(Triangle a b c) = let c' = orthocentre t in do draw white t draw red $ t `ceviansThrough` c' draw red c' demoMittenpunkt :: Demo demoMittenpunkt t@(Triangle a b c) = let mp = mittenpunkt t in do draw white t draw green $ outcircles t draw green $ outcentres t draw cyan $ rays mp (outcentres t) draw cyan $ mp demoIncircle :: Demo demoIncircle t@(Triangle a b c) = let ic@(Circle c r) = incircle t in do draw white t draw yellow $ rays c (corners t) draw yellow $ c draw yellow $ ic demoCircumcircle :: Demo demoCircumcircle t@(Triangle a b c) = let cc@(Circle c r) = circumcircle t in do draw white t draw cyan $ rays c (sidemidpoints t) draw green $ c draw green $ cc demoNinePointCircle :: Demo demoNinePointCircle t@(Triangle a b c) = let o = orthocentre t heights = t `ceviansThrough` o hs = map (\(Line a b) -> b) heights npc@(Circle c r) = ninepointcircle t in do draw white $ t draw cyan $ heights draw blue $ raysMidpoints o (corners t) draw red $ sidemidpoints t draw yellow $ c draw yellow $ npc draw cyan $ hs demoOuterNapoleonTriangle :: Demo demoOuterNapoleonTriangle t = let ots = outerEquilateralTriangles t in do draw red $ ots draw green $ outerNapoleonTriangle t draw white $ t -- isouter is True if you're using outer equilateral triangles, False for inner. demoNapoleonTriangle :: Bool -> Demo demoNapoleonTriangle isouter t = let ets = sideEquilateralTriangles isouter t nt = napoleonTriangle isouter t in do draw red $ ets draw green $ nt draw white $ t demoNagel :: Demo demoNagel t@(Triangle a b c) = let n = nagelpoint t in do draw white $ t draw cyan $ n draw cyan $ t `ceviansThrough` n demoEulerLine :: Demo demoEulerLine t@(Triangle a b c) = let o = orthocentre t hs = cevianIntersects t o cd = centroid t ms = cevianIntersects t cd cc@(Circle ccc ccr) = circumcircle t ex = exeterpoint t npc@(Circle npcc npcr) = ninepointcircle t in do draw (faded white) $ zipWith Line hs ms draw white $ t draw cyan $ t `ceviansThrough` o draw cyan $ rays o (corners t) draw cyan $ o draw green $ t `ceviansThrough` cd draw green $ cd draw blue $ ccc draw blue $ cc draw magenta $ ex draw yellow $ Line o ex draw yellow $ Line ccc ex draw yellow $ npcc draw yellow $ npc --------------------------------------------------------------------------------
clauderichard/Euclaude
Demos.hs
mit
3,906
0
13
1,484
1,270
591
679
115
1
-------------------------------------------------------------------------------- -- | -- Module : Control.Workflow -- Copyright : (c) 2015-2019 Kai Zhang -- License : MIT -- Maintainer : kai@kzhang.org -- Stability : experimental -- Portability : portable -- -- DSL for building computational workflows. Example: -- -- > {-# LANGUAGE TemplateHaskell #-} -- > {-# LANGUAGE OverloadedStrings #-} -- > -- > import Control.Monad.Reader -- > import Control.Concurrent (threadDelay) -- > import Control.Lens -- > import System.Environment -- > import qualified Data.HashMap.Strict as M -- > import Network.Transport.TCP -- > import Data.Proxy (Proxy(..)) -- > -- > import Control.Workflow -- > import Control.Workflow.Coordinator.Local -- > import Control.Workflow.Coordinator.Drmaa -- > -- > s0 :: () -> ReaderT Int IO [Int] -- > s0 = return . const [1..10] -- > -- > s1 :: Int -> ReaderT Int IO Int -- >s1 i = (i*) <$> ask -- > -- > s2 = return . (!!1) -- > s3 = return . (!!2) -- > s4 = return . (!!3) -- > s5 = return . (!!4) -- > s6 (a,b,c,d) = liftIO $ threadDelay 10000000 >> print [a,b,c,d] -- > -- > build "wf" [t| SciFlow Int |] $ do -- > node "S0" 's0 $ return () -- > nodePar "S1" 's1 $ return () -- > ["S0"] ~> "S1" -- > -- > node "S2" 's2 $ memory .= 30 -- > node "S3" 's3 $ memory .= 30 -- > node "S4" 's4 $ nCore .= 4 -- > node "S5" 's5 $ queue .= Just "gpu" -- > ["S0"] ~> "S2" -- > ["S0"] ~> "S3" -- > ["S0"] ~> "S4" -- > ["S0"] ~> "S5" -- > -- > node "S6" 's6 $ return () -- > ["S2", "S3", "S4", "S5"] ~> "S6" -- > main :: IO () -- > main = do -- > let serverAddr = "192.168.0.1" -- > port = 8888 -- > storePath = "sciflow.db" -- > resources = ResourceConfig $ M.fromList -- > [("S6", Resource (Just 2) Nothing Nothing)] -- > [mode] <- getArgs -- > case mode of -- > -- Run on local machine -- > "local" -> withCoordinator (LocalConfig 5) $ \coord -> do -- > Right transport <- createTransport (defaultTCPAddr serverAddr $ show port) -- > defaultTCPParameters -- > withStore storePath $ \store -> -- > runSciFlow coord transport store (ResourceConfig M.empty) 2 wf -- > -- Using the DRMAA backend -- > "drmaa" -> do -- > config <- getDefaultDrmaaConfig ["slave"] -- > withCoordinator config $ \coord -> do -- > Right transport <- createTransport (defaultTCPAddr serverAddr $ show port) -- > defaultTCPParameters -- > withStore storePath $ \store -> -- > runSciFlow coord transport store resources 2 wf -- > -- DRMAA workers -- > "slave" -> startClient (Proxy :: Proxy Drmaa) -- > (mkNodeId serverAddr port) $ _function_table wf -------------------------------------------------------------------------------- module Control.Workflow ( -- * Workflow types SciFlow(..) , ResourceConfig(..) , Resource(..) -- * Construct workflow , Builder , node , nodePar , uNode , doc , nCore , memory , queue , (~>) , path , namespace , build -- * Run workflow , withCoordinator , withStore , startClient , runSciFlow , mkNodeId ) where import Control.Workflow.Types import Control.Workflow.Language import Control.Workflow.Language.TH import Control.Workflow.Interpreter.Exec import Control.Workflow.DataStore import Control.Workflow.Coordinator import Control.Workflow.Utils
kaizhang/SciFlow
SciFlow/src/Control/Workflow.hs
mit
3,631
0
5
998
210
170
40
29
0
{-# LANGUAGE FlexibleContexts #-} {-# LANGUAGE MultiParamTypeClasses #-} {-# LANGUAGE TemplateHaskell #-} {-# LANGUAGE ViewPatterns #-} -- | This module provides abstractions for parallel job processing. module Control.TimeWarp.Manager.Job ( -- * Job data types InterruptType (..) , JobCurator (..) -- * 'JobsState' lenses , jcCounter , jcIsClosed , jcJobs -- * Manager utilities , addManagerAsJob , addSafeThreadJob , addThreadJob , interruptAllJobs , isInterrupted , mkJobCurator , stopAllJobs , unlessInterrupted ) where import Control.Concurrent (forkIO) import Control.Concurrent.STM (atomically, check) import Control.Concurrent.STM.TVar (TVar, newTVarIO, readTVar, readTVarIO, writeTVar) import Control.Lens (at, makeLenses, use, view, (&), (.=), (.~), (<<+=), (<<.=), (?=), (^.)) import Control.Monad (mapM_, unless, void) import Control.Monad.Catch (MonadMask (mask), finally) import Control.Monad.Extra ((&&^)) import Control.Monad.Trans (MonadIO (liftIO)) import Control.Monad.Trans.Control (MonadBaseControl (..)) import Data.Hashable (Hashable) import Data.HashMap.Strict (HashMap) import qualified Data.HashMap.Strict as HM hiding (HashMap) import Serokell.Util.Base (inCurrentContext) import Serokell.Util.Concurrent (modifyTVarS, threadDelay) import System.Wlog (CanLog) import Control.TimeWarp.Timed (Microsecond, MonadTimed, fork_, killThread, myThreadId) -- | Unique identifier of job. newtype JobId = JobId Word deriving (Show, Eq, Num, Hashable) -- | Job killer. newtype JobInterrupter = JobInterrupter { runJobInterrupter :: IO () } -- | Action to mark job as finished newtype MarkJobFinished = MarkJobFinished { runMarker :: IO () } data JobCuratorState = JobCuratorState { -- | @True@ if interrupt had been invoked. Also, when @True@, no job could be added _jcIsClosed :: !Bool -- | 'Map' with currently active jobs , _jcJobs :: !(HashMap JobId JobInterrupter) -- | Total number of allocated jobs ever , _jcCounter :: !JobId } makeLenses ''JobCuratorState -- | Keeps set of jobs. Allows to stop jobs and wait till all of them finish. newtype JobCurator = JobCurator { getJobCurator :: TVar JobCuratorState } -- | Defines way to interrupt all jobs in curator. data InterruptType = Plain -- ^ Just interrupt all jobs | Force -- ^ Interrupt all jobs, and treat them all as completed | WithTimeout !Microsecond !(IO ()) -- ^ Interrupt all jobs in `Plain` was, but if some jobs fail to complete in time, -- interrupt `Force`ly and execute given action. mkJobCurator :: MonadIO m => m JobCurator mkJobCurator = JobCurator <$> (liftIO $ newTVarIO JobCuratorState { _jcIsClosed = False , _jcJobs = mempty , _jcCounter = 0 }) -- | Remembers and starts given action. -- Once `interruptAllJobs` called on this manager, if job is not completed yet, -- `JobInterrupter` is invoked. -- -- Given job *must* invoke given `MarkJobFinished` upon finishing, even if it was -- interrupted. -- -- If curator is already interrupted, action would not start, and `JobInterrupter` -- would be invoked. addJob :: MonadIO m => JobCurator -> JobInterrupter -> (MarkJobFinished -> m ()) -> m () addJob (getJobCurator -> curator) ji@(runJobInterrupter -> interrupter) action = do jidm <- liftIO . atomically $ do st <- readTVar curator let closed = st ^. jcIsClosed if closed then return Nothing else modifyTVarS curator $ do no <- jcCounter <<+= 1 jcJobs . at no ?= ji return $ Just no maybe (liftIO interrupter) (action . markReady) jidm where markReady jid = MarkJobFinished $ atomically $ do st <- readTVar curator writeTVar curator $ st & jcJobs . at jid .~ Nothing -- | Invokes `JobInterrupter`s for all incompleted jobs. -- Has no effect on second call. interruptAllJobs :: MonadIO m => JobCurator -> InterruptType -> m () interruptAllJobs (getJobCurator -> curator) Plain = do jobs <- liftIO . atomically $ modifyTVarS curator $ do wasClosed <- jcIsClosed <<.= True if wasClosed then return mempty else use jcJobs liftIO $ mapM_ runJobInterrupter jobs interruptAllJobs c@(getJobCurator -> curator) Force = do interruptAllJobs c Plain liftIO . atomically $ modifyTVarS curator $ jcJobs .= mempty interruptAllJobs c@(getJobCurator -> curator) (WithTimeout delay onTimeout) = do interruptAllJobs c Plain void $ liftIO . forkIO $ do threadDelay delay done <- HM.null . view jcJobs <$> readTVarIO curator unless done $ liftIO onTimeout >> interruptAllJobs c Force -- | Waits for this manager to get closed and all registered jobs to invoke -- `MaskForJobFinished`. awaitAllJobs :: MonadIO m => JobCurator -> m () awaitAllJobs (getJobCurator -> jc) = liftIO . atomically $ check =<< (view jcIsClosed &&^ (HM.null . view jcJobs)) <$> readTVar jc -- | Interrupts and then awaits for all jobs to complete. stopAllJobs :: MonadIO m => JobCurator -> m () stopAllJobs c = interruptAllJobs c Plain >> awaitAllJobs c -- | Add second manager as a job to first manager. addManagerAsJob :: (MonadIO m, MonadTimed m, MonadBaseControl IO m) => JobCurator -> InterruptType -> JobCurator -> m () addManagerAsJob curator intType managerJob = do interrupter <- inCurrentContext $ interruptAllJobs managerJob intType addJob curator (JobInterrupter interrupter) $ \(runMarker -> ready) -> fork_ $ awaitAllJobs managerJob >> liftIO ready -- | Adds job executing in another thread, where interrupting kills the thread. addThreadJob :: (CanLog m, MonadIO m, MonadMask m, MonadTimed m, MonadBaseControl IO m) => JobCurator -> m () -> m () addThreadJob curator action = mask $ \unmask -> fork_ $ do tid <- myThreadId killer <- inCurrentContext $ killThread tid addJob curator (JobInterrupter killer) $ \(runMarker -> markReady) -> unmask action `finally` liftIO markReady -- | Adds job executing in another thread, interrupting does nothing. -- Usefull then work stops itself on interrupt, and we just need to wait till it fully -- stops. addSafeThreadJob :: (MonadIO m, MonadMask m, MonadTimed m) => JobCurator -> m () -> m () addSafeThreadJob curator action = mask $ \unmask -> fork_ $ addJob curator (JobInterrupter $ return ()) $ \(runMarker -> markReady) -> unmask action `finally` liftIO markReady isInterrupted :: MonadIO m => JobCurator -> m Bool isInterrupted = liftIO . atomically . fmap (view jcIsClosed) . readTVar . getJobCurator unlessInterrupted :: MonadIO m => JobCurator -> m () -> m () unlessInterrupted c a = isInterrupted c >>= flip unless a
serokell/time-warp
src/Control/TimeWarp/Manager/Job.hs
mit
7,469
0
17
2,095
1,664
899
765
-1
-1
{-# LANGUAGE PatternSynonyms, ForeignFunctionInterface, JavaScriptFFI #-} module GHCJS.DOM.JSFFI.Generated.SVGPathSegArcAbs (js_setX, setX, js_getX, getX, js_setY, setY, js_getY, getY, js_setR1, setR1, js_getR1, getR1, js_setR2, setR2, js_getR2, getR2, js_setAngle, setAngle, js_getAngle, getAngle, js_setLargeArcFlag, setLargeArcFlag, js_getLargeArcFlag, getLargeArcFlag, js_setSweepFlag, setSweepFlag, js_getSweepFlag, getSweepFlag, SVGPathSegArcAbs, castToSVGPathSegArcAbs, gTypeSVGPathSegArcAbs) where import Prelude ((.), (==), (>>=), return, IO, Int, Float, Double, Bool(..), Maybe, maybe, fromIntegral, round, fmap, Show, Read, Eq, Ord) import Data.Typeable (Typeable) import GHCJS.Types (JSVal(..), JSString) import GHCJS.Foreign (jsNull) import GHCJS.Foreign.Callback (syncCallback, asyncCallback, syncCallback1, asyncCallback1, syncCallback2, asyncCallback2, OnBlocked(..)) import GHCJS.Marshal (ToJSVal(..), FromJSVal(..)) import GHCJS.Marshal.Pure (PToJSVal(..), PFromJSVal(..)) import Control.Monad.IO.Class (MonadIO(..)) import Data.Int (Int64) import Data.Word (Word, Word64) import GHCJS.DOM.Types import Control.Applicative ((<$>)) import GHCJS.DOM.EventTargetClosures (EventName, unsafeEventName) import GHCJS.DOM.JSFFI.Generated.Enums foreign import javascript unsafe "$1[\"x\"] = $2;" js_setX :: SVGPathSegArcAbs -> Float -> IO () -- | <https://developer.mozilla.org/en-US/docs/Web/API/SVGPathSegArcAbs.x Mozilla SVGPathSegArcAbs.x documentation> setX :: (MonadIO m) => SVGPathSegArcAbs -> Float -> m () setX self val = liftIO (js_setX (self) val) foreign import javascript unsafe "$1[\"x\"]" js_getX :: SVGPathSegArcAbs -> IO Float -- | <https://developer.mozilla.org/en-US/docs/Web/API/SVGPathSegArcAbs.x Mozilla SVGPathSegArcAbs.x documentation> getX :: (MonadIO m) => SVGPathSegArcAbs -> m Float getX self = liftIO (js_getX (self)) foreign import javascript unsafe "$1[\"y\"] = $2;" js_setY :: SVGPathSegArcAbs -> Float -> IO () -- | <https://developer.mozilla.org/en-US/docs/Web/API/SVGPathSegArcAbs.y Mozilla SVGPathSegArcAbs.y documentation> setY :: (MonadIO m) => SVGPathSegArcAbs -> Float -> m () setY self val = liftIO (js_setY (self) val) foreign import javascript unsafe "$1[\"y\"]" js_getY :: SVGPathSegArcAbs -> IO Float -- | <https://developer.mozilla.org/en-US/docs/Web/API/SVGPathSegArcAbs.y Mozilla SVGPathSegArcAbs.y documentation> getY :: (MonadIO m) => SVGPathSegArcAbs -> m Float getY self = liftIO (js_getY (self)) foreign import javascript unsafe "$1[\"r1\"] = $2;" js_setR1 :: SVGPathSegArcAbs -> Float -> IO () -- | <https://developer.mozilla.org/en-US/docs/Web/API/SVGPathSegArcAbs.r1 Mozilla SVGPathSegArcAbs.r1 documentation> setR1 :: (MonadIO m) => SVGPathSegArcAbs -> Float -> m () setR1 self val = liftIO (js_setR1 (self) val) foreign import javascript unsafe "$1[\"r1\"]" js_getR1 :: SVGPathSegArcAbs -> IO Float -- | <https://developer.mozilla.org/en-US/docs/Web/API/SVGPathSegArcAbs.r1 Mozilla SVGPathSegArcAbs.r1 documentation> getR1 :: (MonadIO m) => SVGPathSegArcAbs -> m Float getR1 self = liftIO (js_getR1 (self)) foreign import javascript unsafe "$1[\"r2\"] = $2;" js_setR2 :: SVGPathSegArcAbs -> Float -> IO () -- | <https://developer.mozilla.org/en-US/docs/Web/API/SVGPathSegArcAbs.r2 Mozilla SVGPathSegArcAbs.r2 documentation> setR2 :: (MonadIO m) => SVGPathSegArcAbs -> Float -> m () setR2 self val = liftIO (js_setR2 (self) val) foreign import javascript unsafe "$1[\"r2\"]" js_getR2 :: SVGPathSegArcAbs -> IO Float -- | <https://developer.mozilla.org/en-US/docs/Web/API/SVGPathSegArcAbs.r2 Mozilla SVGPathSegArcAbs.r2 documentation> getR2 :: (MonadIO m) => SVGPathSegArcAbs -> m Float getR2 self = liftIO (js_getR2 (self)) foreign import javascript unsafe "$1[\"angle\"] = $2;" js_setAngle :: SVGPathSegArcAbs -> Float -> IO () -- | <https://developer.mozilla.org/en-US/docs/Web/API/SVGPathSegArcAbs.angle Mozilla SVGPathSegArcAbs.angle documentation> setAngle :: (MonadIO m) => SVGPathSegArcAbs -> Float -> m () setAngle self val = liftIO (js_setAngle (self) val) foreign import javascript unsafe "$1[\"angle\"]" js_getAngle :: SVGPathSegArcAbs -> IO Float -- | <https://developer.mozilla.org/en-US/docs/Web/API/SVGPathSegArcAbs.angle Mozilla SVGPathSegArcAbs.angle documentation> getAngle :: (MonadIO m) => SVGPathSegArcAbs -> m Float getAngle self = liftIO (js_getAngle (self)) foreign import javascript unsafe "$1[\"largeArcFlag\"] = $2;" js_setLargeArcFlag :: SVGPathSegArcAbs -> Bool -> IO () -- | <https://developer.mozilla.org/en-US/docs/Web/API/SVGPathSegArcAbs.largeArcFlag Mozilla SVGPathSegArcAbs.largeArcFlag documentation> setLargeArcFlag :: (MonadIO m) => SVGPathSegArcAbs -> Bool -> m () setLargeArcFlag self val = liftIO (js_setLargeArcFlag (self) val) foreign import javascript unsafe "($1[\"largeArcFlag\"] ? 1 : 0)" js_getLargeArcFlag :: SVGPathSegArcAbs -> IO Bool -- | <https://developer.mozilla.org/en-US/docs/Web/API/SVGPathSegArcAbs.largeArcFlag Mozilla SVGPathSegArcAbs.largeArcFlag documentation> getLargeArcFlag :: (MonadIO m) => SVGPathSegArcAbs -> m Bool getLargeArcFlag self = liftIO (js_getLargeArcFlag (self)) foreign import javascript unsafe "$1[\"sweepFlag\"] = $2;" js_setSweepFlag :: SVGPathSegArcAbs -> Bool -> IO () -- | <https://developer.mozilla.org/en-US/docs/Web/API/SVGPathSegArcAbs.sweepFlag Mozilla SVGPathSegArcAbs.sweepFlag documentation> setSweepFlag :: (MonadIO m) => SVGPathSegArcAbs -> Bool -> m () setSweepFlag self val = liftIO (js_setSweepFlag (self) val) foreign import javascript unsafe "($1[\"sweepFlag\"] ? 1 : 0)" js_getSweepFlag :: SVGPathSegArcAbs -> IO Bool -- | <https://developer.mozilla.org/en-US/docs/Web/API/SVGPathSegArcAbs.sweepFlag Mozilla SVGPathSegArcAbs.sweepFlag documentation> getSweepFlag :: (MonadIO m) => SVGPathSegArcAbs -> m Bool getSweepFlag self = liftIO (js_getSweepFlag (self))
manyoo/ghcjs-dom
ghcjs-dom-jsffi/src/GHCJS/DOM/JSFFI/Generated/SVGPathSegArcAbs.hs
mit
6,051
98
8
841
1,374
763
611
78
1
{-# OPTIONS_HADDOCK show-extensions #-} {-# LANGUAGE FlexibleContexts #-} -- | -- -- Module : EventProbability.NaiveBayes.Weka -- Description : NaiveBayes for Weka data. -- License : MIT -- -- NaiveBayes learing and classification for Weka data. -- module EventProbability.NaiveBayes.Weka ( -- * Prepare wekaEntry2Event -- * Execute , buildCaches , classifyWekaData , ClassifyResult -- * Test , learnAndTest , Filename , runWekaLearnAndTest ) where import Cache import EventProbability import EventProbability.Cache import EventProbability.NaiveBayes import WekaData import qualified Data.Set as Set import Data.Maybe ( fromMaybe, maybeToList ) import Data.List ( find ) import Data.Typeable import Control.Arrow ( (&&&) ) import Control.Monad ( when ) ----------------------------------------------------------------------------- instance AtomicEvent WekaVal where eventName (WVal attr _) = EventName $ wekaAttributeName attr eventDomain (WVal attr@(WekaAttrNom _ dom) _) = Set.fromList $ map (uncurry WVal . (const attr &&& id)) dom ----------------------------------------------------------------------------- -- | Build an 'Event' from a 'WekaEntry'. wekaEntry2Event :: (Show WekaVal) => WekaEntry -> Event wekaEntry2Event (WEntry vals) = compositeEvent vals ----------------------------------------------------------------------------- -- | Create initial 'EventCaches' and update counts of the 'EventCountCache'. buildCaches :: (Show WekaVal) => RawWekaData -> IO (EventCaches IO) buildCaches rawData = do cc <- newIOCache pc <- newIOCache cpc <- newIOCache updateCountCache cc events return $ EvCaches cc pc cpc where events = map wekaEntry2Event $ wekaData2Sparse rawData ----------------------------------------------------------------------------- -- | Classification result of an 'Event'. type ClassifyResult = (Event, Maybe (EvAtom, Prob)) -- | Classifies given 'RawWekaData' by given attribute name using 'EventCaches'. classifyWekaData :: (Show WekaVal, NaiveBayesCondProb (EventCaches IO) IO) => EventCaches IO -> RawWekaData -- ^ data to classify. -> String -- ^ attribute to clasify. -> IO [ClassifyResult] classifyWekaData caches rawData className = sequence $ do ev <- events return $ do c <- classifyEvent caches cAtom ev return (ev, c) where classifyBy = fromMaybe (error $ errStr ++ show className) $ find ((==) className . wekaAttributeName) $ rwdAttrs rawData errStr = "couldn't find an attribute with name " cAtom = EvAtom $ WVal classifyBy undefined events = map wekaEntry2Event $ wekaData2Sparse rawData ----------------------------------------------------------------------------- -- | Teaches a /Naive Bayes/ classifier (updates the caches) with the -- /learn/ data and tests the classification quality with the /test/ data. learnAndTest :: ( Show WekaVal, Show WekaEntry , NaiveBayesCondProb (EventCaches IO) IO) => RawWekaData -- ^ /learn/ data. -> RawWekaData -- ^ /test/ data. -> String -- ^ attribute to classify. -> Bool -- ^ show test results. -> IO (EventCaches IO, [ClassifyResult]) learnAndTest learnData testData className showRes = do caches <- buildCaches learnData testClassify <- classifyWekaData caches testData className -- for showRes let err = "class '" ++ className ++ "' not found in " let getClass e@(WEntry set) = find (\(WVal a _)-> className == wekaAttributeName a) $ Set.toList set let expectedRes = map getClass $ wekaData2Sparse testData let notClassified ev exp = do putStrLn $ "\n[?]" ++ show ev ++ " wasn't classified, expected " ++ show exp return False let compareClassified ev exp (EvAtom c, p) = case cast exp of Just exp' | exp' == c -> do putStrLn strOk return True _ -> do putStrLn strFail return False where strOk = "\n[+]\t" ++ show ev ++ " was classified correcly (" ++ show c ++ ") with probability " ++ show p strFail = "\n[-]\t" ++ show ev ++ " was classified incorrecly: expected " ++ show exp ++ ", got " ++ show c ++ "(with probability " ++ show p ++ ")" when showRes $ do results <- sequence $ do ((ev, mbClass), mbExp) <- zip testClassify expectedRes let compare exp = maybe (notClassified ev exp) (compareClassified ev exp) mbClass maybe [] (return . compare) mbExp let ok = length $ filter id results putStrLn $ replicate 30 '=' putStrLn $ show ok ++ "\t classified correctly" putStrLn $ show (length results - ok) ++ "\t classified incorrectly" return (caches, testClassify) ----------------------------------------------------------------------------- type Filename = String -- | Calls 'learnAndTest' after extracting 'RawWekaData' from the given files. runWekaLearnAndTest :: ( Show WekaVal, Show WekaEntry , NaiveBayesCondProb (EventCaches IO) IO) => Filename -- ^ path to a weka data file with /learn/ data. -> Filename -- ^ path to a weka data file with /test/ data. -> String -- ^ attribute to classify. -> Bool -- ^ show test results. -> IO (EventCaches IO, [ClassifyResult]) runWekaLearnAndTest learnFile testFile className showRes = do learnData <- readWekaData learnFile testData <- readWekaData testFile learnAndTest learnData testData className showRes -----------------------------------------------------------------------------
fehu/min-dat--naive-bayes
src/EventProbability/NaiveBayes/Weka.hs
mit
6,125
0
20
1,726
1,276
647
629
100
2
{-# LANGUAGE DeriveGeneric #-} module U.Codebase.Kind where import U.Util.Hashable (Hashable) import qualified U.Util.Hashable as Hashable import GHC.Generics (Generic) data Kind = Star | Arrow Kind Kind deriving (Eq,Ord,Read,Show,Generic) instance Hashable Kind where tokens k = case k of Star -> [Hashable.Tag 0] Arrow k1 k2 -> (Hashable.Tag 1 : Hashable.tokens k1) ++ Hashable.tokens k2
unisonweb/platform
codebase2/codebase/U/Codebase/Kind.hs
mit
404
0
13
68
146
80
66
10
0
-- Disguised sequences (II) -- https://www.codewars.com/kata/56fe17fcc25bf3e19a000292 module Codewars.G964.Disguised2 where u1 :: Integer -> Integer -> Integer u1 n p = p * succ n v1 :: Integer -> Integer -> Integer v1 n p = p * succ (2 * n) uEff :: Integer -> Integer -> Integer uEff = u1 vEff :: Integer -> Integer -> Integer vEff = v1
gafiatulin/codewars
src/6 kyu/Disguised2.hs
mit
343
0
8
66
112
61
51
9
1
{-# LANGUAGE ScopedTypeVariables #-} {-# LANGUAGE DeriveGeneric #-} module Network.OnDemandSSHTunnel ( SSHTunnel(..), Config(..), prettyConfig, setupConfiguration, setupOnDemandTunnel ) where import System.Process import Control.Monad import Control.Concurrent import Control.Exception import qualified Control.Exception as CE import Network.Socket import System.Random import System.IO import qualified Data.ByteString as BS import Text.Printf (printf) import System.Mem import Text.PrettyPrint.GenericPretty data SSHTunnel = SSHTunnel Int String [String] deriving (Show, Read, Generic) newtype Config = Config [SSHTunnel] deriving (Show, Read, Generic) instance Out SSHTunnel instance Out Config prettyConfig :: Config -> String prettyConfig = pretty setupConfiguration :: Config -> IO () setupConfiguration (Config tuns) = forM_ tuns $ forkIO . setupOnDemandTunnel setupOnDemandTunnel :: SSHTunnel -> IO () setupOnDemandTunnel (SSHTunnel listenPort targetHostPort sshArgs) = do sock <- socket AF_INET Stream defaultProtocol setSocketOption sock ReuseAddr 1 addr <- inet_addr "127.0.0.1" bindSocket sock $ SockAddrInet (fromIntegral listenPort) addr listen sock 5 putStrLn $ "listening for connections on port " ++ show listenPort forever $ do (conn, connaddr) <- accept sock hConn <- socketToHandle conn ReadWriteMode putStrLn $ "connection accepted: " ++ show connaddr forkIO $ do tunnelPort :: Int <- randomRIO (50000, 55000) handleConnection hConn tunnelPort targetHostPort sshArgs handleConnection :: Handle -> Int -> String -> [String] -> IO () handleConnection hConn tunnelPort targetHostPort sshArgs = do (_, _, _, p) <- createProcess (proc "ssh" $ ["-v"] ++ [ "-L" ++ show tunnelPort ++ ":" ++ targetHostPort, "-n"] ++ sshArgs) finally forwardToTunnel $ do putStrLn "terminating ssh" terminateProcess p _ <- waitForProcess p putStrLn "ssh exited" performGC -- a good moment to finalize handles where forwardToTunnel = do hTunn <- tryTunnelConnection tunnelPort 15 _ <- forkIO $ forward "hTunn->hConn" hTunn hConn forward "hConn->hTunn" hConn hTunn forward desc h1 h2 = forever $ do bs <- BS.hGetSome h1 4096 putStrLn $ printf "%s %d bytes" desc (BS.length bs) if BS.null bs then error "no more data to tunnel" else do BS.hPut h2 bs hFlush h2 tryTunnelConnection :: Int -> Int -> IO Handle tryTunnelConnection _ 0 = error "can't connect to tunnel" tryTunnelConnection port attempts = do threadDelay (1*1000*1000) tunn <- socket AF_INET Stream defaultProtocol addr <- inet_addr "127.0.0.1" CE.catch (do connect tunn $ SockAddrInet (fromIntegral port) addr hTunn <- socketToHandle tunn ReadWriteMode putStrLn "got hTunn" return hTunn) (\(_::SomeException) -> do sClose tunn tryTunnelConnection port (attempts-1))
crackleware/on-demand-ssh-tunnel
src/Network/OnDemandSSHTunnel.hs
mit
3,125
0
16
774
883
433
450
83
2
{-# LANGUAGE OverloadedStrings #-} {-# LANGUAGE RecordWildCards #-} {-# LANGUAGE StrictData #-} {-# LANGUAGE TupleSections #-} -- | http://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-properties-redshift-cluster-loggingproperties.html module Stratosphere.ResourceProperties.RedshiftClusterLoggingProperties where import Stratosphere.ResourceImports -- | Full data type definition for RedshiftClusterLoggingProperties. See -- 'redshiftClusterLoggingProperties' for a more convenient constructor. data RedshiftClusterLoggingProperties = RedshiftClusterLoggingProperties { _redshiftClusterLoggingPropertiesBucketName :: Val Text , _redshiftClusterLoggingPropertiesS3KeyPrefix :: Maybe (Val Text) } deriving (Show, Eq) instance ToJSON RedshiftClusterLoggingProperties where toJSON RedshiftClusterLoggingProperties{..} = object $ catMaybes [ (Just . ("BucketName",) . toJSON) _redshiftClusterLoggingPropertiesBucketName , fmap (("S3KeyPrefix",) . toJSON) _redshiftClusterLoggingPropertiesS3KeyPrefix ] -- | Constructor for 'RedshiftClusterLoggingProperties' containing required -- fields as arguments. redshiftClusterLoggingProperties :: Val Text -- ^ 'rclpBucketName' -> RedshiftClusterLoggingProperties redshiftClusterLoggingProperties bucketNamearg = RedshiftClusterLoggingProperties { _redshiftClusterLoggingPropertiesBucketName = bucketNamearg , _redshiftClusterLoggingPropertiesS3KeyPrefix = Nothing } -- | http://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-properties-redshift-cluster-loggingproperties.html#cfn-redshift-cluster-loggingproperties-bucketname rclpBucketName :: Lens' RedshiftClusterLoggingProperties (Val Text) rclpBucketName = lens _redshiftClusterLoggingPropertiesBucketName (\s a -> s { _redshiftClusterLoggingPropertiesBucketName = a }) -- | http://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-properties-redshift-cluster-loggingproperties.html#cfn-redshift-cluster-loggingproperties-s3keyprefix rclpS3KeyPrefix :: Lens' RedshiftClusterLoggingProperties (Maybe (Val Text)) rclpS3KeyPrefix = lens _redshiftClusterLoggingPropertiesS3KeyPrefix (\s a -> s { _redshiftClusterLoggingPropertiesS3KeyPrefix = a })
frontrowed/stratosphere
library-gen/Stratosphere/ResourceProperties/RedshiftClusterLoggingProperties.hs
mit
2,221
0
13
212
265
151
114
28
1
{-# LANGUAGE OverloadedStrings #-} module Network.API.Mandrill.Messages where import Network.API.Mandrill.Utils import Network.API.Mandrill.Response import Network.API.Mandrill.Types -- | Send a new transactional message through Mandrill send :: (MonadIO m) => Message -> MessageConfig -> MandrillT m (Either ApiError [DeliveryStatus]) send m c = performRequest "/messages/send.json" $ [ "message" .= m , "ip_pool" .= _conf_ip_pool c , "async" .= _conf_async c , "send_at" .= _conf_send_at c ] -- | Send a new transactional message through Mandrill using a template sendTmpl :: (MonadIO m) => Message -> MessageConfig -> Name -> [Content] -> MandrillT m (Either ApiError [DeliveryStatus]) sendTmpl m c t tc = performRequest "/messages/send-template.json" $ [ "template_name" .= t , "template_content" .= tc , "message" .= m , "ip_pool" .= _conf_ip_pool c , "async" .= _conf_async c , "send_at" .= _conf_send_at c ] -- | Search recently sent messages and optionally narrow by date -- range, tags, senders, and API keys. If no date range is specified, -- results within the last 7 days are returned. This method may be called -- up to 20 times per minute. If you need the data more often, -- you can use /messages/info.json to get the information for a single message, -- or webhooks to push activity to your own application for querying. search :: (MonadIO m) => Query -> From -> Thru -> [Tag] -> [Email] -> [ApiKey] -> Count -> MandrillT m (Either ApiError [SearchResult]) search q f t ts es ks c = performRequest "/messages/search.json" $ [ "query" .= q , "date_from" .= f , "date_to" .= t , "tags" .= ts , "senders" .= es , "api_keys" .= ks , "limit" .= c ] -- | Search the content of recently sent messages and return the aggregated -- hourly stats for matching messages searchTimeSeries :: (MonadIO m ) => Query -> From -> Thru -> [Tag] -> [Email] -> MandrillT m (Either ApiError [Stat]) searchTimeSeries q f t ts es = performRequest "/messages/search-time-series.json" $ [ "query" .= q , "date_from" .= f , "date_to" .= t , "tags" .= ts , "senders" .= es ] -- | Get the information for a single recently sent message info :: (MonadIO m) => MessageId -> MandrillT m (Either ApiError SearchResult) info mid = performRequest "/messages/info.json" ["id" .= mid] -- | Get the full content of a recently sent message content :: (MonadIO m ) => MessageId -> MandrillT m (Either ApiError Message) content mid = performRequest "/messages/content.json" ["id" .= mid] -- | Parse the full MIME document for an email message, returning the content -- of the message broken into its constituent pieces parse :: (MonadIO m) => RawMessage -> MandrillT m (Either ApiError Message) parse raw = performRequest "/messages/parse.json" ["raw_message" .= raw] -- | Take a raw MIME document for a message, and send it exactly as if it -- were sent through Mandrill's SMTP servers sendRaw :: (MonadIO m) => RawMessage -> Email -> Name -> [Recipient] -> MessageConfig -> MandrillT m (Either ApiError [DeliveryStatus]) sendRaw raw e n to c = performRequest "/messages/send-raw.json" $ [ "raw_message" .= raw , "from_email" .= e , "from_name" .= n , "to" .= to , "async" .= _conf_async c , "ip_pool" .= _conf_ip_pool c , "send_at" .= _conf_send_at c ] -- | Queries your scheduled emails by sender or recipient, or both. listScheduled :: (MonadIO m) => Email -> MandrillT m (Either ApiError [Scheduled]) listScheduled e = performRequest "/messages/list-scheduled.json" ["to" .= e] -- | Cancels a scheduled email. cancelScheduled :: (MonadIO m) => ScheduledId -> MandrillT m (Either ApiError Scheduled) cancelScheduled i = performRequest "/messages/cancel-scheduled.json" ["id" .= i] -- | Reschedules a scheduled email. reschedule :: (MonadIO m) => ScheduledId -> TimeStamp -> MandrillT m (Either ApiError Scheduled) reschedule i s = performRequest "/messages/reschedule.json" $ [ "id" .= i , "send_at" .= s ]
krgn/hamdrill
src/Network/API/Mandrill/Messages.hs
mit
4,831
0
16
1,575
977
525
452
110
1
-- Making our own types and typeclasses module Shapes ( Point(..) , Shape(..) , surface , nudge , baseCircle , baseRect ) where data Point = Point Float Float deriving (Show) data Shape = Circle Point Float | Rectangle Point Point deriving (Show) surface :: Shape -> Float surface (Circle _ r) = pi * r ^ 2 surface (Rectangle (Point x1 y1) (Point x2 y2)) = (abs $ x2 - x1) * (abs $ y2 - y1) nudge :: Shape -> Float -> Float -> Shape nudge (Circle (Point x y) r) a b = Circle (Point (x+a) (y+b)) r nudge (Rectangle (Point x1 y1) (Point x2 y2)) a b = Rectangle (Point (x1+a) (y1+b)) (Point (x2+a) (y2+b)) baseCircle :: Float -> Shape baseCircle r = Circle (Point 0 0) r baseRect :: Float -> Float -> Shape baseRect width height = Rectangle (Point 0 0) (Point width height) -- record syntax, fields get functions that get the field data Person = Person { firstName :: String , lastName :: String , age :: Int , height :: Float , phoneNumber :: String , flavor :: String } deriving (Show) data Car = Car { company :: String , model :: String , year :: Int } deriving (Show) -- its a very strong convention in haskell to never add typeclass -- constraints in data declarations.
autocorr/lyah
chap8.hs
mit
1,237
0
9
291
487
269
218
35
1
-- ------------------------------------------------------------------------------------- -- Author: Sourabh S Joshi (cbrghostrider); Copyright - All rights reserved. -- For email, run on linux (perl v5.8.5): -- perl -e 'print pack "H*","736f75726162682e732e6a6f73686940676d61696c2e636f6d0a"' -- ------------------------------------------------------------------------------------- import Data.List import Data.Ord minDiffFrom :: Int -> [Int] -> Int minDiffFrom fromme = minimum . map ((abs) . (\x -> fromme - x)) solveProblem :: Int -> Int -> [Int] -> Int solveProblem p q ns = let ns' = sort ns pcand = minDiffFrom p ns' qcand = minDiffFrom q ns' mids = filter (\(x, dx)-> (x >= p) && (x <= q)) . reverse . snd . foldl' (\a@(ln, ls) nn -> (nn, ((ln+nn) `div` 2, (nn-ln) `div` 2):ls)) (head ns', []) $ tail ns' allCands = (p, pcand):(q, qcand):mids maxdelta = maximum . map snd $ allCands chosen = head . sort . map fst . filter ((== maxdelta) . snd) $ allCands in chosen main :: IO () main = do _ <- getLine nsstr <- getLine pqstr <- getLine let ns = map read . words $ nsstr [p, q] = map read . words $ pqstr ans = solveProblem p q ns putStrLn $ show ans
cbrghostrider/Hacking
HackerRank/Algorithms/Greedy/minimax.hs
mit
1,287
0
19
322
451
242
209
24
1
module Compiling where import Control.Monad.Writer import System.Process import System.Exit import Text.Printf import Data.List.Split data GC = JGC | Stub data TDir = NoTDir | TDirAt FilePath data CompilerFlags = CF { tdir :: TDir , entryPoint :: Maybe String , cCompiler :: String , extraCFlags :: [String] , includes :: [String] , hsCompiler :: String , ignoreCache :: Bool , debug :: Bool , extraHaskellFlags :: [String] , toC :: Bool , cPreprocessor :: [(PreprocessorFlag, Maybe String)] , hsPreprocessor :: Maybe [(PreprocessorFlag, Maybe String)] , gc :: GC } type PreprocessorFlag = String preprocessorFlags :: [(PreprocessorFlag, Maybe String)] -> [String] preprocessorFlags = map $ \pair -> case pair of (flag, Nothing) -> "-D"++flag (flag, Just val) -> "-D"++flag++"="++val whenJust :: (Monad m) => Maybe a -> (a -> m ()) -> m () whenJust Nothing _ = return () whenJust (Just a) f = f a fileToModuleName :: String -> String fileToModuleName = takeWhile (/= '.') compilerFlagsToHaskell :: FilePath -> Maybe FilePath -> CompilerFlags -> [String] compilerFlagsToHaskell input output flags = snd $ runWriter $ do case gc flags of JGC -> return () Stub -> tell ["-fjgc-stub"] whenJust (entryPoint flags) $ \main -> tell ["--main=" ++ fileToModuleName input ++ "." ++ main] when (ignoreCache flags) $ tell ["--ignore-cache"] whenJust (hsPreprocessor flags) $ \ppFlags -> do tell ["-fcpp"] tell $ preprocessorFlags ppFlags tell $ extraHaskellFlags flags case tdir flags of NoTDir -> return () TDirAt dir -> tell ["--tdir=" ++ dir] --TDirRandom -> return () -- TODO implement this at some point when (toC flags) (tell ["-C"]) whenJust output (\filename -> tell ["--output=" ++ filename]) tell [input] includeDirectives :: [String] -> [String] includeDirectives = map ("-I" ++) compilerFlagsToC :: [FilePath] -> Maybe FilePath -> CompilerFlags -> [String] compilerFlagsToC inputs output flags = snd $ runWriter $ do case gc flags of JGC -> tell ["-D_JHC_GC=_JHC_GC_JGC"] Stub -> tell ["-D_JHC_GC=_JHC_GC_JGC", "-D_JHC_GC_JGC_STUB"] tell $ includeDirectives $ includes flags tell $ extraCFlags flags when (debug flags) $ tell ["-g"] tell $ preprocessorFlags $ cPreprocessor flags whenJust output (\filename -> tell ["-o " ++ filename]) tell inputs runHaskellCompiler :: FilePath -> Maybe FilePath -> CompilerFlags -> IO(ExitCode,String,String) runHaskellCompiler input output flags = do printf "Command: %s %s\n" compiler (unwords haskellFlags) readProcessWithExitCode compiler haskellFlags "" where compiler = hsCompiler flags haskellFlags = compilerFlagsToHaskell input output flags runCCompiler :: [FilePath] -> Maybe FilePath -> CompilerFlags -> IO (ExitCode, String, String) runCCompiler inputs output flags = do printf "Command: %s %s\n" compiler (unwords cFlags) readProcessWithExitCode compiler cFlags "" where compiler = cCompiler flags cFlags = compilerFlagsToC inputs output flags report :: ExitCode -> String -> String -> String -> IO ExitCode report exitCode input out err = do case exitCode of ExitFailure i -> do printf "Error compiling inputs %s\n" input printf "Error code %d\n" i printf "STDOUT:\n%s\n" out printf "STDERR:\n%s\n" err ExitSuccess -> return () return exitCode compileHs :: FilePath -> Maybe FilePath -> CompilerFlags -> IO ExitCode compileHs input output flags = do (exitCode, out, err) <- runHaskellCompiler input output flags report exitCode input out err runLinker :: [FilePath] -> Maybe FilePath -> IO (ExitCode, String, String) runLinker files output = readProcessWithExitCode "llvm-link" (files ++ out) "" where out = case output of Nothing -> [] Just file -> ["-o=" ++ file] replaceExtension :: String -> String replaceExtension path = takeWhile (/= '.') (last $ splitOn "/" path) ++ ".o" compileC :: [FilePath] -> Maybe FilePath -> CompilerFlags -> IO ExitCode compileC inputs output flags = do (exitCode, out, err) <- runCCompiler inputs Nothing flags _ <- report exitCode (show inputs) out err case exitCode of ExitFailure _ -> return exitCode ExitSuccess -> do (exitCode', out', err') <- runLinker (map replaceExtension inputs) output _ <- report exitCode' (show $ map replaceExtension inputs) out' err' _ <- system "rm -rf *.o" return exitCode'
m-alvarez/scher-run
Compiling.hs
mit
4,836
0
16
1,273
1,531
773
758
112
3
{-# LANGUAGE ScopedTypeVariables #-} {-# LANGUAGE TypeFamilies #-} {-# LANGUAGE RankNTypes #-} {-# LANGUAGE NamedFieldPuns #-} -- A simple structure for a pool of resources. -- Each managed resource has an integer index that identifies the 'slot' for the given resource; -- when the resource is idle, this slot holds a Nothing; when it's currently used it holds the -- content of the resource module Data.Pool ( Pool , Id -- * Accessors , busy , idles , slots -- * Query , null , size -- * Creation , empty , idling , fromBusy -- * Modification , takeSlot , takeSlot_ , takeSlots , takeSlots_ , releaseSlot , reserve , growBy ) where import Prelude hiding (null) import Data.Tuple (swap) import Data.List (mapAccumL) import qualified Data.IntMap.Strict as IM import qualified Data.IntSet as IS type Id = Int -- elems hold the elements in use -- idles contains the currently unoccupied ids (but issued in the past) -- nexts contains the lazy list of all the future ids to issue -- therefore, keysSet elems \cap idles = \0 data Pool a = Pool (IM.IntMap a) -- the busy / occupied cells IS.IntSet -- the idle cells Id -- the base of the remaining id deriving (Show, Eq, Ord) -- TODO: Show instance, Eq instance..? busy :: Pool a -> IM.IntMap a busy (Pool busy _ _) = busy idles :: Pool a -> IS.IntSet idles (Pool _ idles _) = idles -- all the slots in the pool, those busy contain a 'Just a'; whereas the idle ones have a Nothing slots :: Pool a -> IM.IntMap (Maybe a) slots (Pool busy idles _) = IM.union (Just <$> busy) (IM.fromSet (const Nothing) idles) null :: Pool a -> Bool null = (&&) <$> IM.null . busy <*> IS.null . idles size :: Pool a -> Int size = (+) <$> IM.size . busy <*> IS.size . idles empty :: Pool a empty = Pool IM.empty IS.empty 0 idling :: Int -> Pool a idling c = Pool IM.empty (IS.fromList [0..c-1]) c -- | call takeSlots on the elements with an empty starting pool fromBusy :: [a] -> Pool a fromBusy = flip takeSlots_ empty takeSlot_ :: a -> Pool a -> Pool a takeSlot_ a = snd . takeSlot a takeSlot :: a -> Pool a -> (Id, Pool a) takeSlot a (Pool busy idles next) | IS.null idles = (next, Pool (IM.insert next a busy) idles (next+1)) | otherwise = let (k, idles') = IS.deleteFindMin idles in (k, Pool (IM.insert k a busy) idles' next) -- insert all elements into the pool, starting from the left takeSlots :: [a] -> Pool a -> ([Id], Pool a) takeSlots l p = swap $ mapAccumL (\p a -> let (k, p') = takeSlot a p in (p', k)) p l takeSlots_ :: [a] -> Pool a -> Pool a takeSlots_ l = snd . takeSlots l -- | reserve a certain minimum size reserve :: Int -> Pool a -> Pool a reserve c p | left > 0 = growBy left p | otherwise = p where left = c - size p growBy :: Int -> Pool a -> Pool a growBy s (Pool busy idles next) = Pool busy (foldr IS.insert idles $ take s [next..]) (next+s) releaseSlot :: Id -> Pool a -> Pool a releaseSlot k p@(Pool busy idles next) | IM.member k busy = Pool (IM.delete k busy) (IS.insert k idles) next | otherwise = p
lynnard/reflex-cocos2d
src/Data/Pool.hs
mit
3,086
0
13
707
1,025
542
483
73
1
module Filter.Hyperref (hyperref) where import Text.Pandoc.Definition import Text.Pandoc.Walk (walk) tex :: String -> Inline tex = RawInline (Format "tex") f :: Inline -> Inline f x@(Link attr is (('#':id), title)) = Span nullAttr $ [ tex $ "\\hyperref[" ++ id ++ "]{" ] ++ is ++ [ tex "}" ] f x = x hyperref :: Pandoc -> Pandoc hyperref = walk f
Thhethssmuz/ppp
src/Filter/Hyperref.hs
mit
353
0
11
69
155
85
70
11
1
#!/usr/bin/env runhaskell {-# LANGUAGE DataKinds #-} {-# LANGUAGE LambdaCase #-} {-# LANGUAGE OverloadedStrings #-} {-# LANGUAGE UnicodeSyntax #-} {-# LANGUAGE ViewPatterns #-} import Data.Foldable (foldMap) import Data.Maybe (fromMaybe) import Data.Monoid import Control.Lens import qualified Data.Aeson.Lens as L import Network.StackExchange main ∷ IO () main = ask >>= print . foldMap count ask ∷ IO [SE Badge] ask = askSE $ badgesOnUsers [972985] <> site "stackoverflow" <> key "Lhg6xe5d5BvNK*C0S8jijA((" count ∷ SE Badge → (Sum Int, Sum Int, Sum Int) count (view (from se) → x) = (\l → set l (x ^. L.key "award_count" . to (Sum . fromMaybe 0)) mempty) $ case x ^. L.key "rank" . L.asText of Just "bronze" → _1 Just "silver" → _2 Just "gold" → _3 _ → error "badge rank isn't bronze/silver/gold"
supki/libstackexchange
examples/badges-watcher.hs
mit
870
0
14
180
277
144
133
-1
-1
-- -- -- ------------------ -- Exercise 12.41. ------------------ -- -- -- module E'12'41 where
pascal-knodel/haskell-craft
_/links/E'12'41.hs
mit
106
0
2
24
13
12
1
1
0
module Foundation where import Prelude import Yesod import Yesod.Static import Yesod.Auth import Yesod.Auth.BrowserId import Yesod.Auth.GoogleEmail import Yesod.Default.Config import Yesod.Default.Util (addStaticContentExternal) import Network.HTTP.Conduit (Manager) import qualified Settings import Settings.Development (development) import qualified Database.Persist import Database.Persist.Sql (SqlPersistT) import Settings.StaticFiles import Settings (widgetFile, Extra (..)) import Model import Text.Jasmine (minifym) import Text.Hamlet (hamletFile) import Yesod.Core.Types (Logger) import Data.Text -- | The site argument for your application. This can be a good place to -- keep settings and values requiring initialization before your application -- starts running, such as database connections. Every handler will have -- access to the data present here. data App = App { settings :: AppConfig DefaultEnv Extra , getStatic :: Static -- ^ Settings for static file serving. , connPool :: Database.Persist.PersistConfigPool Settings.PersistConf -- ^ Database connection pool. , httpManager :: Manager , persistConfig :: Settings.PersistConf , appLogger :: Logger } -- Set up i18n messages. See the message folder. mkMessage "App" "messages" "en" -- This is where we define all of the routes in our application. For a full -- explanation of the syntax, please see: -- http://www.yesodweb.com/book/routing-and-handlers -- -- Note that this is really half the story; in Application.hs, mkYesodDispatch -- generates the rest of the code. Please see the linked documentation for an -- explanation for this split. mkYesodData "App" $(parseRoutesFile "config/routes") type Form x = Html -> MForm (HandlerT App IO) (FormResult x, Widget) -- Please see the documentation for the Yesod typeclass. There are a number -- of settings which can be configured by overriding methods here. instance Yesod App where approot = ApprootMaster $ appRoot . settings -- Store session data on the client in encrypted cookies, -- default session idle timeout is 120 minutes makeSessionBackend _ = fmap Just $ defaultClientSessionBackend (120 * 60) -- 120 minutes "config/client_session_key.aes" defaultLayout widget = do master <- getYesod mmsg <- getMessage -- We break up the default layout into two components: -- default-layout is the contents of the body tag, and -- default-layout-wrapper is the entire page. Since the final -- value passed to hamletToRepHtml cannot be a widget, this allows -- you to use normal widget features in default-layout. pc <- widgetToPageContent $ do $(combineStylesheets 'StaticR [ css_normalize_css , css_bootstrap_css ]) $(widgetFile "default-layout") giveUrlRenderer $(hamletFile "templates/default-layout-wrapper.hamlet") -- This is done to provide an optimization for serving static files from -- a separate domain. Please see the staticRoot setting in Settings.hs urlRenderOverride y (StaticR s) = Just $ uncurry (joinPath y (Settings.staticRoot $ settings y)) $ renderRoute s urlRenderOverride _ _ = Nothing -- The page to be redirected to when authentication is required. authRoute _ = Just $ AuthR LoginR -- This function creates static content files in the static folder -- and names them based on a hash of their content. This allows -- expiration dates to be set far in the future without worry of -- users receiving stale content. addStaticContent = addStaticContentExternal minifym genFileName Settings.staticDir (StaticR . flip StaticRoute []) where -- Generate a unique filename based on the content itself genFileName lbs | development = "autogen-" ++ base64md5 lbs | otherwise = base64md5 lbs -- Place Javascript at bottom of the body tag so the rest of the page loads first jsLoader _ = BottomOfBody -- What messages should be logged. The following includes all messages when -- in development, and warnings and errors in production. shouldLog _ _source level = development || level == LevelWarn || level == LevelError makeLogger = return . appLogger -- How to run database actions. instance YesodPersist App where type YesodPersistBackend App = SqlPersistT runDB = defaultRunDB persistConfig connPool instance YesodPersistRunner App where getDBRunner = defaultGetDBRunner connPool instance YesodAuth App where type AuthId App = UserId -- Where to send a user after successful login loginDest _ = HomeR -- Where to send a user after logout logoutDest _ = HomeR getAuthId creds = runDB $ do x <- getBy $ UniqueUser $ credsIdent creds case x of Just (Entity uid _) -> return $ Just uid Nothing -> do fmap Just $ insert $ User (credsIdent creds) Nothing -- You can add other plugins like BrowserID, email or OAuth here authPlugins _ = [authBrowserId def, authGoogleEmail] authHttpManager = httpManager -- This instance is required to use forms. You can modify renderMessage to -- achieve customized and internationalized form validation messages. instance RenderMessage App FormMessage where renderMessage _ _ = defaultFormMessage -- | Get the 'Extra' value, used to hold data from the settings.yml file. getExtra :: Handler Extra getExtra = fmap (appExtra . settings) getYesod -- Note: previous versions of the scaffolding included a deliver function to -- send emails. Unfortunately, there are too many different options for us to -- give a reasonable default. Instead, the information is available on the -- wiki: -- -- https://github.com/yesodweb/yesod/wiki/Sending-email
ardumont/yesod-lab
Foundation.hs
gpl-2.0
5,888
0
17
1,275
859
471
388
-1
-1
{-| Path-related helper functions. -} {- Copyright (C) 2012 Google Inc. This program is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with this program; if not, write to the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. -} module Ganeti.Path ( dataDir , runDir , logDir , socketDir , defaultLuxiSocket , defaultQuerySocket , confdHmacKey , clusterConfFile , nodedCertFile , queueDir , jobQueueSerialFile , jobQueueArchiveSubDir , instanceReasonDir , getInstReasonFilename ) where import System.FilePath import System.Posix.Env (getEnvDefault) import AutoConf -- | Simple helper to concat two paths. pjoin :: IO String -> String -> IO String pjoin a b = do a' <- a return $ a' </> b -- | Returns the root directory, which can be either the real root or -- the virtual root. getRootDir :: IO FilePath getRootDir = getEnvDefault "GANETI_ROOTDIR" "" -- | Prefixes a path with the current root directory. addNodePrefix :: FilePath -> IO FilePath addNodePrefix path = do root <- getRootDir return $ root ++ path -- | Directory for data. dataDir :: IO FilePath dataDir = addNodePrefix $ AutoConf.localstatedir </> "lib" </> "ganeti" -- | Helper for building on top of dataDir (internal). dataDirP :: FilePath -> IO FilePath dataDirP = (dataDir `pjoin`) -- | Directory for runtime files. runDir :: IO FilePath runDir = addNodePrefix $ AutoConf.localstatedir </> "run" </> "ganeti" -- | Directory for log files. logDir :: IO FilePath logDir = addNodePrefix $ AutoConf.localstatedir </> "log" </> "ganeti" -- | Directory for Unix sockets. socketDir :: IO FilePath socketDir = runDir `pjoin` "socket" -- | The default LUXI socket path. defaultLuxiSocket :: IO FilePath defaultLuxiSocket = socketDir `pjoin` "ganeti-master" -- | The default LUXI socket for queries. defaultQuerySocket :: IO FilePath defaultQuerySocket = socketDir `pjoin` "ganeti-query" -- | Path to file containing confd's HMAC key. confdHmacKey :: IO FilePath confdHmacKey = dataDirP "hmac.key" -- | Path to cluster configuration file. clusterConfFile :: IO FilePath clusterConfFile = dataDirP "config.data" -- | Path to the noded certificate. nodedCertFile :: IO FilePath nodedCertFile = dataDirP "server.pem" -- | Job queue directory. queueDir :: IO FilePath queueDir = dataDirP "queue" -- | Job queue serial file. jobQueueSerialFile :: IO FilePath jobQueueSerialFile = queueDir `pjoin` "serial" -- | Job queue archive directory. jobQueueArchiveSubDir :: FilePath jobQueueArchiveSubDir = "archive" -- | Directory containing the reason trails for the last change of status of -- instances. instanceReasonDir :: IO FilePath instanceReasonDir = runDir `pjoin` "instance-reason" -- | The path of the file containing the reason trail for an instance, given the -- instance name. getInstReasonFilename :: String -> IO FilePath getInstReasonFilename instName = instanceReasonDir `pjoin` instName
vladimir-ipatov/ganeti
src/Ganeti/Path.hs
gpl-2.0
3,446
0
8
584
494
280
214
58
1
module Handler.Home where import Import getHomeR :: Handler Html getHomeR = defaultLayout $ do setTitle "Welcome!" $(widgetFile "homepage")
cblp/tasknight-dashboard
frontend/Handler/Home.hs
gpl-3.0
162
0
10
40
42
21
21
7
1
{-# LANGUAGE NamedFieldPuns #-} {-# LANGUAGE TypeOperators #-} import Data.Monoid ((<>)) import Data.Traversable (for) import Distribution.Package (PackageId) import Distribution.Text (simpleParse) import Options.Applicative (execParser, fullDesc, help, helper, info, long, metavar, optional, progDesc, strOption) import Stack.Offline (Config(..), Packages(..), Snapshot, createPack) data Options = Options { options_resolver :: Snapshot , options_packagesFromFile :: Maybe FilePath , options_outFile :: FilePath } deriving Show main :: IO () main = parseOptions >>= resolveConfig >>= createPack where parseOptions = execParser $ info (helper <*> options) programDescription programDescription = mconcat [ fullDesc , progDesc "Create portable environment sufficient to build your stack project offline" ] options = Options <$> strOptionM [long "resolver", metavar "SNAPSHOT", help "Snapshot to use"] <*> optional ( strOptionM [ long "packages-from" , metavar "FILE" , help $ unwords [ "Read packages list from file." , "Each line of the file must be in pkgid format (<name>-<version>)." ] ] ) <*> strOptionM [ long "tgz", metavar "FILE", help "Result tar.gz file" ] strOptionM = strOption . mconcat resolveConfig :: Options -> IO Config resolveConfig Options{options_resolver, options_packagesFromFile, options_outFile} = do config_packages <- case options_packagesFromFile of Nothing -> pure PackagesFromShapshot Just filePath -> Packages <$> loadPackageList filePath pure Config {config_snapshot = options_resolver, config_packages, config_outFile = options_outFile} loadPackageList :: FilePath -> IO [PackageId] loadPackageList filePath = do contents <- readFile filePath for (zip [1 :: Int ..] $ lines contents) $ \(n, line) -> case simpleParse line of Just pkgid -> pure pkgid Nothing -> error $ filePath <> ":" <> show n <> ": cannot parse PackageId"
cblp/stack-offline
app/Main.hs
gpl-3.0
2,252
0
15
645
519
280
239
45
2
{-# LANGUAGE OverloadedStrings #-} {-| Module : Web.CampBX Description : An API Library for the CampBX Bitcoin Market Copyright : (c) Pavan Rikhi, 2014 License : GPL-3 Maintainer : pavan@sleepanarchy.com Stability : experimental Portability : POSIX CampBX is a library for interacting with the API for CampBX, a Bitcoin Trading Market. Actions are run in the 'CampBX' Monad, which should be supplied with a 'CampBXConfig'. You will need to override the 'bxUser' and 'bxPass' Config fields to use the 'Authorized Endpoints'. The following actions are currently supported: * Retrieving the Market Depth, Ticker, Pending Orders and Account Balances * Depositing & Withdrawing Funds from the CampBX Account * Placing Quick & Advanced Buy/Sell Orders * Canceling Orders -} module Web.CampBX ( -- * CampBX Monad CampBX , runCampBX , defaultCampBXConfig -- * CampBX API Functions -- ** Anonymous Endpoints , getDepth , getTicker -- ** Authorized Endpoints , getWallet , getDepositAddress , getPendingOrders , sendBTC , placeQuickBuy , placeBuy , placeQuickSell , placeSell , cancelBuyOrder , cancelSellOrder -- * CampBX Types -- ** Currency & Offer Types , USDAmount , BTCAmount , BTCPrice , BTCAddress , Ask(..) , Bid(..) -- ** Response Types , Ticker(..) , Depth(..) , Wallet(..) , DepositAddress(..) , TransferResponse(..) -- ** Order-related Types , OrderList(..) , Order(..) , OrderType(..) , PriceType(..) , FillType(..) , Margin(..) , DarkPool(..) -- ** API-related Types , CampBXConfig(..) , APIStatus(..) ) where import qualified Data.ByteString.Char8 as BC import Web.CampBX.Client import Web.CampBX.Types -- | Retrieves all the Buy & Sell Offers on the Market getDepth :: CampBX Depth getDepth = queryEndPoint GetDepth [] -- | Retrieves the Current Market Ticker getTicker :: CampBX Ticker getTicker = queryEndPoint GetTicker [] -- | Retrieves the User's Total, Liquid and Margin Account Funds getWallet :: CampBX Wallet getWallet = queryEndPoint GetFunds [] -- | Retrieves a Bitcoin Address for Depositing Funds into the CampBX -- Account getDepositAddress :: CampBX DepositAddress getDepositAddress = queryEndPoint GetBTCAddr [] -- | Retrieves the Account's Open Orders getPendingOrders :: CampBX OrderList getPendingOrders = queryEndPoint GetOrders [] -- | Sends Bitcoins from the CampBX Account to the Specified Address sendBTC :: BTCAddress -> BTCAmount -> CampBX Integer sendBTC address quantity = queryEndPoint SendBTC [ ("BTCTo", BC.pack address) , ("BTCAmt", BC.pack $ show quantity) ] -- | Places a Limit Order to Buy a Quantity of BTC at or Below a Given Price placeQuickBuy :: BTCAmount -> BTCPrice -> CampBX (APIStatus Int) placeQuickBuy = placeQuickOrder "QuickBuy" -- | Places a Limit Order to Sell a Quantity of BTC at or Above a Given Price placeQuickSell :: BTCAmount -> BTCPrice -> CampBX (APIStatus Int) placeQuickSell = placeQuickOrder "QuickSell" -- | Places a Limit Order of the Supplied TradeMode placeQuickOrder :: BC.ByteString -> BTCAmount -> BTCPrice -> CampBX (APIStatus Int) placeQuickOrder tm quantity price = queryEndPoint TradeEnter [ ("TradeMode", tm) , ("Quantity", BC.pack $ show quantity) , ("Price", BC.pack $ show price) ] -- | Places an Advanced Buy Order With an Optional FillType, Dark Pool and -- Expiration placeBuy :: BTCAmount -> BTCPrice -> Maybe FillType -> Maybe DarkPool -> Maybe String -> CampBX (APIStatus Int) placeBuy = placeOrder "AdvancedBuy" -- | Places an Advanced Sell Order With an Optional FillType, Dark Pool and -- Expiration placeSell :: BTCAmount -> BTCPrice -> Maybe FillType -> Maybe DarkPool -> Maybe String -> CampBX (APIStatus Int) placeSell = placeOrder "AdvancedSell" -- | Places an Advanced Order with an Optional FillType, DarkPool and -- Expiration Date placeOrder :: BC.ByteString -> BTCAmount -> BTCPrice -> Maybe FillType -> Maybe DarkPool -> Maybe String -> CampBX (APIStatus Int) placeOrder tm q p Nothing dp e = placeOrder tm q p (Just Incremental) dp e placeOrder tm q p ft Nothing e = placeOrder tm q p ft (Just No) e placeOrder tm q p ft dp Nothing = placeOrder tm q p ft dp (Just "") placeOrder tm q p (Just ft) (Just dp) (Just e) = queryEndPoint TradeAdvanced [ ("TradeMode", tm) , ("Quantity", BC.pack $ show q) , ("Price", BC.pack $ show p) , ("FillType", BC.pack $ show ft) , ("DarkPool", BC.pack $ show dp) , ("Expiry", BC.pack e) ] -- | Cancels the Buy Order with the Given Order ID cancelBuyOrder :: Int -> CampBX (APIStatus Int) cancelBuyOrder orderID = queryEndPoint TradeCancel [ ("Type", "Buy") , ("OrderID", BC.pack $ show orderID) ] -- | Cancels the Sell Order with the Given Order ID cancelSellOrder :: Int -> CampBX (APIStatus Int) cancelSellOrder orderID = queryEndPoint TradeCancel [ ("Type", "Sell") , ("OrderID", BC.pack $ show orderID) ]
prikhi/campbx-haskell
src/Web/CampBX.hs
gpl-3.0
5,909
0
13
1,911
1,054
589
465
91
1
{-# OPTIONS -fno-warn-orphans #-} -- -- Copyright (c) 2005-8 Don Stewart - http://www.cse.unsw.edu.au/~dons -- -- This program is free software; you can redistribute it and/or -- modify it under the terms of the GNU General Public License as -- published by the Free Software Foundation; either version 2 of -- the License, or (at your option) any later version. -- -- This program is distributed in the hope that it will be useful, -- but WITHOUT ANY WARRANTY; without even the implied warranty of -- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -- General Public License for more details. -- -- You should have received a copy of the GNU General Public License -- along with this program; if not, write to the Free Software -- Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA -- 02111-1307, USA. -- -- -- functions for manipulating file trees -- module Tree where import Codec.Compression.GZip import qualified Data.ByteString.Char8 as P import qualified Data.ByteString.Lazy as L import FastIO import Syntax (Mode (..)) import Data.Binary import Data.Array import Data.Char (toLower) import Data.List (foldl', groupBy, sort, sortBy) import Data.Maybe (catMaybes) import Control.Exception import Control.Monad (liftM) import System.Directory (Permissions (readable)) import System.IO (hPutStrLn, stderr) type FilePathP = P.ByteString -- | A filesystem hierarchy is flattened to just the end nodes type DirArray = Array Int Dir -- | The complete list of .mp3 files type FileArray = Array Int File -- | A directory entry is the directory name, and a list of bound -- indicies into the Files array. data Dir = Dir { dname :: !FilePathP -- ^ directory name , dsize :: !Int -- ^ number of file entries , dlo :: !Int -- ^ index of first entry , dhi :: !Int } -- ^ index of last entry -- Most data is allocated in this structure data File = File { fbase :: !FilePathP -- ^ basename of file , fdir :: !Int } -- ^ index of Dir entry -- -- | Given the start directories, populate the dirs and files arrays -- buildTree :: [FilePathP] -> IO (DirArray, FileArray) buildTree fs = do (os,dirs) <- partition fs -- note we will lose the ordering of files given on cmd line. let loop xs | seq xs False = undefined -- strictify loop [] = return [] loop (a:xs) = do (m,ds) <- expandDir a ms <- loop $! ds ++ xs -- add to work list return $! m : ms ms' <- liftM catMaybes $! loop dirs let extras = merge . doOrphans $ os ms = ms' ++ extras let (_,n,dirls,filels) = foldl' make (0,0,[],[]) ms dirsArray = listArray (0,length dirls - 1) (reverse dirls) fileArray = listArray (0, n-1) (reverse filels) dirsArray `seq` fileArray `seq` return $ (dirsArray, fileArray) -- | Create nodes based on dirname for orphan files on cmdline doOrphans :: [FilePathP] -> [(FilePathP, [FilePathP])] doOrphans [] = [] doOrphans (f:xs) = (dirnameP f, [basenameP f]) : doOrphans xs -- | Merge entries with the same root node into a single node merge :: [(FilePathP, [FilePathP])] -> [(FilePathP, [FilePathP])] merge [] = [] merge xs = let xs' = sortBy (\a b -> fst a `compare` fst b) xs xs''= groupBy (\a b -> fst a == fst b) xs' in catMaybes $ map flatten xs'' where flatten :: [(FilePathP,[FilePathP])] -> Maybe (FilePathP, [FilePathP]) flatten [] = Nothing -- can't happen flatten (x:ys) = let d = fst x in Just (d, snd x ++ concatMap snd ys) -- | fold builder, for generating Dirs and Files make :: (Int,Int,[Dir],[File]) -> (FilePathP,[FilePathP]) -> (Int,Int,[Dir],[File]) make (i,n,acc1,acc2) (d,fs) = case listToDir n d fs of (dir,n') -> case map makeFile fs of fs' -> (i+1, n', dir:acc1, (reverse fs') ++ acc2) where makeFile f = File (basenameP f) i ------------------------------------------------------------------------ -- | Expand a single directory into a maybe a pair of the dir name and any files -- Return any extra directories to search in -- -- Assumes no evil sym links -- expandDir :: FilePathP -> IO (Maybe (FilePathP, [FilePathP]), [FilePathP]) expandDir f | seq f False = undefined -- stricitfy expandDir f = do ls_raw <- handle (\(SomeException e) -> hPutStrLn stderr (show e) >> return []) $ packedGetDirectoryContents f let ls = map (\s -> P.intercalate (P.singleton '/') [f,s]) . sort . filter validFiles $! ls_raw ls `seq` return () (fs',ds) <- partition ls let fs = filter onlyMp3s fs' v = if null fs then Nothing else Just (f,fs) return (v,ds) where notEdge p = p /= dot && p /= dotdot validFiles p = notEdge p onlyMp3s p = mp3 == (P.map toLower . P.drop (P.length p -3) $ p) mp3 = P.pack "mp3" dot = P.pack "." dotdot = P.pack ".." -- -- | Given an the next index into the files array, a directory name, and -- a list of files in that dir, build a Dir and return the next index -- into the array -- listToDir :: Int -> FilePathP -> [FilePathP] -> (Dir, Int) listToDir n d fs = let dir = Dir { dname = packedFileNameEndClean d , dsize = len , dlo = n , dhi = n + len - 1 } in (dir, n') where len = length fs n' = n + len -- | break a list of file paths into a pair of subliests corresponding -- to the paths that point to files and to directories. partition :: [FilePathP] -> IO ([FilePathP], [FilePathP]) partition xs | seq xs False = undefined -- how to make `partition' strict partition [] = return ([],[]) partition (a:xs) = do (fs,ds) <- partition xs x <- doesFileExist a if x then do y <- getPermissions a >>= return . readable return $! if y then (a:fs, ds) else (fs, ds) else return (fs, a:ds) ------------------------------------------------------------------------ -- -- And some more Binary instances -- {- instance Binary a => Binary (Array Int a) where put arr = do put (bounds arr) mapM_ put (elems arr) get = do ((x,y) :: (Int,Int)) <- get (els :: [a]) <- sequence $ take (y+1) $ repeat get return $! listArray (x,y) els -} instance Binary File where put (File nm i) = put nm >> put i get = do nm <- get i <- get return (File nm i) instance Binary Dir where put (Dir nm sz lo hi) = put nm >> put sz >> put lo >> put hi get = do nm <- get sz <- get lo <- get hi <- get return (Dir nm sz lo hi) instance Binary Mode where put = put . fromEnum get = liftM toEnum get -- How we write everything out instance Binary SerialT where put st = do put (ser_farr st, ser_darr st) put (ser_indx st) put (ser_mode st) get = do (a,b)<- get i <- get m <- get return $ SerialT { ser_farr = a ,ser_darr = b ,ser_indx = i ,ser_mode = m } ----------------------------------------------------------------------- -- | Wrap up the values we're going to dump to disk data SerialT = SerialT { ser_farr :: FileArray, ser_darr :: DirArray, ser_indx :: Int, ser_mode :: Mode } writeTree :: FilePath -> SerialT -> IO () writeTree f s = L.writeFile f (compress (encode s)) readTree :: FilePath -> IO SerialT readTree f = do s <- L.readFile f return (decode (decompress s))
danplubell/hmp3-phoenix
src/Tree.hs
gpl-3.0
7,962
0
19
2,416
2,171
1,175
996
148
3
import Streaming import qualified Streaming.Prelude as S import Streaming.Extensible aa = do yield_ () yield_ "July" yield_ () yield_ "August" yield_ () yield_ "September" yield_ () yield_ "October"
michaelt/extensible-streaming
examples/Examples.hs
gpl-3.0
216
0
8
45
78
34
44
12
1
{-# LANGUAGE TemplateHaskell #-} {-# OPTIONS_GHC -fno-warn-unused-matches #-} module Gonimo.Server.State where import Control.Concurrent.STM (STM, TVar, modifyTVar, readTVar, retry) import Control.Lens import Data.Map.Strict (Map) import qualified Data.Map.Strict as M import Data.Set (Set) import qualified Data.Set as S import Data.Text (Text) import Gonimo.Server.DbEntities (ClientId, FamilyId) import Gonimo.Server.Types (Secret) type ChannelSecrets = Map ClientId (ClientId, Secret) type ChannelData = Map Secret (ClientId, Text) data FamilyOnlineState = FamilyOnlineState { _channelSecrets :: ChannelSecrets , _channelData :: ChannelData , _onlineMembers :: Set ClientId } $(makeLenses ''FamilyOnlineState) type FamilyMap = Map FamilyId (TVar FamilyOnlineState) type OnlineState = TVar FamilyMap putSecret :: Secret -> ClientId -> ClientId -> TVar FamilyOnlineState -> STM () -- | putSecret inserts possibly overwrites putSecret secret fromId toId familyStateVar = do t <- _channelSecrets <$> readTVar familyStateVar if toId `M.member` t then retry else modifyTVar familyStateVar (channelSecrets %~ toId `M.insert` (fromId, secret)) receiveSecret :: ClientId -> TVar FamilyOnlineState -> STM (Maybe (ClientId, Secret)) receiveSecret toId familyStateVar = do t <- _channelSecrets <$> readTVar familyStateVar case toId `M.lookup` t of Nothing -> return Nothing Just cs -> do modifyTVar familyStateVar (channelSecrets %~ M.delete toId) return $ Just cs --deleteSecret :: ClientId -> ChannelSecrets -> ChannelSecrets --deleteSecret toId = undefined -- ChannelSecrets . M.delete toId . fetch onlineMember :: ClientId -> TVar FamilyOnlineState -> STM Bool onlineMember cid familyStateVar = do familyState <- readTVar familyStateVar return $ cid `S.member` (familyState^.onlineMembers) putData :: Text -> Secret -> ClientId -> TVar FamilyOnlineState -> STM () -- | putSecret inserts possibly overwrites putData txt secret fromId familyStateVar = do t <- _channelData <$> readTVar familyStateVar if secret `M.member` t then retry else modifyTVar familyStateVar (channelData %~ secret `M.insert` (fromId, txt)) receieveData :: Secret -> TVar FamilyOnlineState -> STM (Maybe (ClientId, Text)) receieveData secret familyStateVar = do t <- _channelData <$> readTVar familyStateVar case secret `M.lookup` t of Nothing -> return Nothing Just cs -> do modifyTVar familyStateVar (channelData %~ M.delete secret) return $ Just cs
charringer/gonimo-back
src/Gonimo/Server/State.hs
agpl-3.0
2,820
0
16
717
717
382
335
51
2
{-# LANGUAGE LambdaCase #-} {-# OPTIONS_GHC -F -pgmF scripts/local-htfpp #-} module UniqTests2 (uniqTests2Main) where import Test.Framework import Control.Concurrent test_fast :: IO () test_fast = let l = flip map [True, False] $ \case True -> 1 False -> 2 in putStrLn (show l) uniqTests2Main args = htfMainWithArgs args htf_thisModulesTests
skogsbaer/HTF
tests/real-bbt/UniqTests2.hs
lgpl-2.1
445
0
12
152
98
52
46
13
2
module DDF.Meta.Dual where newtype Dual l r = Dual {runDual :: (l, r)} instance Eq l => Eq (Dual l r) where (Dual (l, _)) == (Dual (r, _)) = l == r instance Ord l => Ord (Dual l r) where (Dual (l, _)) `compare` (Dual (r, _)) = l `compare` r dualOrig (Dual (l, _)) = l dualDiff (Dual (_, r)) = r mkDual l r = Dual (l, r)
ThoughtWorksInc/DeepDarkFantasy
DDF/Meta/Dual.hs
apache-2.0
329
0
9
81
215
120
95
9
1
<?xml version='1.0' encoding='UTF-8'?> <!DOCTYPE helpset PUBLIC "-//Sun Microsystems Inc.//DTD JavaHelp HelpSet Version 2.0//EN" "http://java.sun.com/products/javahelp/helpset_2_0.dtd"> <helpset version="2.0" xml:lang="en-GB"> <title>Context Alert Filters | ZAP Extension</title> <maps> <homeID>top</homeID> <mapref location="map.jhm"/> </maps> <view> <name>TOC</name> <label>Contents</label> <type>org.zaproxy.zap.extension.help.ZapTocView</type> <data>toc.xml</data> </view> <view> <name>Index</name> <label>Index</label> <type>javax.help.IndexView</type> <data>index.xml</data> </view> <view> <name>Search</name> <label>Search</label> <type>javax.help.SearchView</type> <data engine="com.sun.java.help.search.DefaultSearchEngine"> JavaHelpSearch </data> </view> <view> <name>Favorites</name> <label>Favorites</label> <type>javax.help.FavoritesView</type> </view> </helpset>
0xkasun/security-tools
src/org/zaproxy/zap/extension/alertFilters/resources/help/helpset.hs
apache-2.0
995
82
67
173
425
217
208
-1
-1
{-# LANGUAGE CPP, OverloadedStrings, PackageImports, ScopedTypeVariables #-} -- | In which Hawk's command-line arguments are structured into a `HawkSpec`. module System.Console.Hawk.Args.Parse (parseArgs) where import Prelude hiding (fail) #if MIN_VERSION_base(4,12,0) import Control.Monad.Fail (fail) #else import Prelude (fail) #endif import Data.Char (isSpace) import Data.Maybe import "mtl" Control.Monad.Trans import Control.Monad.Trans.OptionParser import Control.Monad.Trans.Uncertain import qualified System.Console.Hawk.Args.Option as Option import System.Console.Hawk.Args.Option (HawkOption, options) import System.Console.Hawk.Args.Spec import System.Console.Hawk.Context.Dir -- $setup -- >>> let testP parser = runUncertainIO . runOptionParserT options parser -- | (record processor, field processor) type CommonProcessors = (Processor, Processor) -- | Extract '-D' and '-d'. We perform this step separately because those two -- delimiters are used by both the input and output specs. -- -- >>> let test = testP commonProcessors -- -- >>> test [] -- (SeparateOn (Delimiter "\n"),SeparateOn Whitespace) -- -- >>> test ["-D\\n", "-d\\t"] -- (SeparateOn (Delimiter "\n"),SeparateOn (Delimiter "\t")) -- -- >>> test ["-D|", "-d,"] -- (SeparateOn (Delimiter "|"),SeparateOn (Delimiter ",")) -- -- >>> test ["-D", "-d"] -- (DoNotSeparate,DoNotSeparate) commonProcessors :: forall m. (Functor m, Monad m) => OptionParserT HawkOption m CommonProcessors commonProcessors = do r <- consumeProcessor Option.RecordDelimiter defaultRecordSeparator f <- consumeProcessor Option.FieldDelimiter defaultFieldSeparator return (r, f) where consumeProcessor :: HawkOption -> Separator -> OptionParserT HawkOption m Processor consumeProcessor opt def = fromMaybe (SeparateOn def) <$> consumeLast opt (Option.processorConsumer) -- | The input delimiters have already been parsed, but we still need to -- interpret them and to determine the input source. -- -- >>> :{ -- let test = testP $ do { c <- commonProcessors -- ; _ <- consumeExtra stringConsumer -- skip expr -- ; i <- inputSpec c -- ; lift $ print $ inputSource i -- ; lift $ print $ inputFormat i -- } -- :} -- -- >>> test [] -- UseStdin -- Records (Delimiter "\n") (Fields Whitespace) -- -- >>> test ["-d", "-a", "L.reverse"] -- UseStdin -- Records (Delimiter "\n") RawRecord -- -- >>> test ["-D", "-a", "B.reverse"] -- UseStdin -- RawStream -- -- >>> test ["-d:", "-m", "L.head", "/etc/passwd"] -- InputFile "/etc/passwd" -- Records (Delimiter "\n") (Fields (Delimiter ":")) inputSpec :: (Functor m, Monad m) => CommonProcessors -> OptionParserT HawkOption m InputSpec inputSpec (rProc, fProc) = InputSpec <$> source <*> format where source = do r <- consumeExtra stringConsumer return $ case r of Nothing -> UseStdin Just f -> InputFile f format = return streamFormat streamFormat = case rProc of DoNotSeparate -> RawStream SeparateOn rSep -> Records rSep recordFormat recordFormat = case fProc of DoNotSeparate -> RawRecord SeparateOn fSep -> Fields fSep -- | The output delimiters take priority over the input delimiters, regardless -- of the order in which they appear. -- -- >>> :{ -- let test = testP $ do { c <- commonProcessors -- ; o <- outputSpec c -- ; let OutputFormat r f = outputFormat o -- ; lift $ print $ outputSink o -- ; lift $ print (r, f) -- } -- :} -- -- >>> test [] -- UseStdout -- ("\n"," ") -- -- >>> test ["-D;", "-d", "-a", "L.reverse"] -- UseStdout -- (";"," ") -- -- >>> test ["-o\t", "-d,", "-O|"] -- UseStdout -- ("|","\t") outputSpec :: forall m. (Functor m, Monad m) => CommonProcessors -> OptionParserT HawkOption m OutputSpec outputSpec (r, f) = OutputSpec <$> sink <*> format where sink :: OptionParserT HawkOption m OutputSink sink = return UseStdout format :: OptionParserT HawkOption m OutputFormat format = OutputFormat <$> record <*> field record, field :: OptionParserT HawkOption m Delimiter record = fmap (fromMaybe r') $ consumeLast Option.OutputRecordDelimiter $ fromMaybe "" <$> optionalConsumer Option.delimiterConsumer field = fmap (fromMaybe f') $ consumeLast Option.OutputFieldDelimiter $ fromMaybe "" <$> optionalConsumer Option.delimiterConsumer r', f' :: Delimiter r' = fromProcessor defaultRecordDelimiter r f' = fromProcessor defaultFieldDelimiter f -- | The information we need in order to evaluate a user expression: -- the expression itself, and the context in which it should be evaluated. -- In Hawk, that context is the user prelude. -- -- >>> :{ -- let test = testP $ do { e <- exprSpec -- ; lift $ print $ untypedExpr e -- ; lift $ print $ userContextDirectory (contextSpec e) -- } -- :} -- -- >>> test [] -- error: missing user expression -- *** Exception: ExitFailure 1 -- -- >>> test [""] -- error: user expression cannot be empty -- *** Exception: ExitFailure 1 -- -- >>> test ["-D;", "-d", "-a", "L.reverse","-c","somedir"] -- "L.reverse" -- "somedir" exprSpec :: (Functor m, MonadIO m) => OptionParserT HawkOption m ExprSpec exprSpec = ExprSpec <$> (ContextSpec <$> contextDir) <*> expr where contextDir = do maybeDir <- consumeLast Option.ContextDirectory stringConsumer case maybeDir of Nothing -> liftIO findContextFromCurrDirOrDefault Just dir -> return dir expr = do r <- consumeExtra stringConsumer case r of Just e -> if all isSpace e then fail "user expression cannot be empty" else return e Nothing -> fail "missing user expression" -- | Parse command-line arguments to construct a `HawkSpec`. -- -- TODO: complain if some arguments are unused (except perhaps "-d" and "-D"). -- -- >>> :{ -- let test args = do { spec <- runUncertainIO $ parseArgs args -- ; case spec of -- Help -> putStrLn "Help" -- Version -> putStrLn "Version" -- Eval e o -> putStrLn "Eval" >> print (untypedExpr e) >> print (recordDelimiter (outputFormat o), fieldDelimiter (outputFormat o)) -- Apply e i o -> putStrLn "Apply" >> print (untypedExpr e, inputSource i) >> print (inputFormat i) >> print (recordDelimiter (outputFormat o), fieldDelimiter (outputFormat o)) -- Map e i o -> putStrLn "Map" >> print (untypedExpr e, inputSource i) >> print (inputFormat i) >> print (recordDelimiter (outputFormat o), fieldDelimiter (outputFormat o)) -- } -- :} -- -- >>> test [] -- Help -- -- >>> test ["--help"] -- Help -- -- >>> test ["--version"] -- Version -- -- >>> test ["-d\\t", "L.head"] -- Eval -- "L.head" -- ("\n","\t") -- -- >>> test ["-D\r\n", "-d\\t", "-m", "L.head"] -- Map -- ("L.head",UseStdin) -- Records (Delimiter "\r\n") (Fields (Delimiter "\t")) -- ("\r\n","\t") -- -- >>> test ["-D", "-O\n", "-m", "L.head", "file.in"] -- Map -- ("L.head",InputFile "file.in") -- RawStream -- ("\n"," ") parseArgs :: (Functor m,MonadIO m) => [String] -> UncertainT m HawkSpec parseArgs [] = return Help parseArgs args = runOptionParserT options parser args where parser = do lift $ return () -- silence a warning cmd <- fromMaybe eval <$> consumeExclusive assoc c <- commonProcessors cmd c assoc = [ (Option.Help, help) , (Option.Version, version) , (Option.Apply, apply) , (Option.Map, map') ] help, version, eval, apply, map' :: (Functor m,MonadIO m) => CommonProcessors -> OptionParserT HawkOption m HawkSpec help _ = return Help version _ = return Version eval c = Eval <$> exprSpec <*> outputSpec c apply c = Apply <$> exprSpec <*> inputSpec c <*> outputSpec c map' c = Map <$> exprSpec <*> inputSpec c <*> outputSpec c
gelisam/hawk
src/System/Console/Hawk/Args/Parse.hs
apache-2.0
8,499
0
13
2,300
1,226
706
520
86
4
module Main where primes = filter prime [2..] where prime x = null [d | d <- [2..x-1], rem x d == 0] primes' = filterPrimes [2..] where filterPrimes (p:xs) = p : filterPrimes [x | x <- xs, rem x p /= 0]
frankiesardo/seven-languages-in-seven-weeks
src/main/haskell/day2/primes.hs
apache-2.0
253
0
12
94
122
63
59
5
1
module Main where import qualified Language.P4 as P4 import CodeGen import Options.Applicative parseAndFixup :: FilePath -> String -> Either String P4.Program parseAndFixup fileName input = do ast <- P4.parseProgram fileName input P4.fixupSEMAs ast astDump :: FilePath -> IO () astDump file = do contents <- readFile file case parseAndFixup file contents of Left err -> print err Right ast -> print ast compile :: FilePath -> IO () compile file = do contents <- readFile file case parseAndFixup file contents of Left err -> print err Right ast -> writeFile "softswitch.c" (ppCgen ast) data Cmd = AstDump FilePath | Compile FilePath deriving (Show) cmd :: Parser Cmd cmd = subparser ( command "ast-dump" (info cmdAstDump (progDesc "Parse and dump AST.")) <> command "compile" (info cmdCompile (progDesc "Compile P4 program.")) ) cmdAstDump :: Parser Cmd cmdAstDump = AstDump <$> argument str (metavar "PROGRAM") cmdCompile :: Parser Cmd cmdCompile = Compile <$> argument str (metavar "PROGRAM") run :: Cmd -> IO () run (AstDump file) = astDump file run (Compile file) = compile file main :: IO () main = execParser opts >>= run where opts = info (cmd <**> helper) idm
brooksbp/P4
src/compiler/Main.hs
apache-2.0
1,243
0
12
266
443
216
227
39
2
module Notification where import DBus.Notify import DBus.Client (connectSession) giveYourEeysABreak :: IO () giveYourEeysABreak = do client <- connectSession do let startNote = appNote { summary="Mike Says" , body=(Just $ Text "Give your eyes a break") } notification <- notify client startNote putStrLn "Done" where appNote = blankNote { appName="Mike" }
akalyaev/mike-on-linux
src/Notification.hs
apache-2.0
408
0
16
102
110
58
52
12
1
{-# LANGUAGE TemplateHaskell, OverloadedStrings, TupleSections #-} {-# OPTIONS_GHC -fno-warn-orphans #-} module Data.QcIntegrated where import Data.Integrated hiding (tests) import Test.QuickCheck import Test.QuickCheck.All import Test.QuickCheck.Monadic import Test.QuickCheck.Property.Comb import qualified Data.Property as P import qualified Data.Suite.Ops as O import Control.Lens hiding (lens, mapping) import Control.Applicative import Data.Integrated.Partition import Data.Integrated.TestModule import Control.Monad.State import Util.Invariants import Data.String.Builder import Data.ModulePath hiding (toPath) import Data.Path import Arbitrary.Properties import Arbitrary.TestModule (toGenerated) import Arbitrary.FS import qualified Data.Set as S import qualified Data.Map as M import qualified Filesystem.Path.CurrentOS as FS import qualified Data.List as L import Prelude hiding (FilePath) import System.Directory import Env main :: IO () main = void $quickCheckAll prop_toPartition :: Diff -> Property prop_toPartition diff = runInvariants (diff, toPartition diff) inv_toPartition where -- TODO/FIXME cleanup remove diff type inv_toPartition :: Invariants (Diff, Partition) inv_toPartition = do satcomp (\t -> (incoming . fst $ t, snd t)) $ eqCoverageInv "union of [equal, displacing, added] form improper subset of incoming" [equal, displacing, added] satcomp (fromPart . snd) disjointSetsInv satcomp (\t -> (fst t, added . snd $ t)) addedInv satcomp (\t -> (incoming . fst $ t, displacing . snd $ t)) displacingInv satcomp (\t -> (fst t, equal . snd $ t)) equalInv where fromPart p = toSets p [ ("equal", S.fromList . M.keys . equal), ("displacing", S.fromList . M.keys . displacing), ("added", S.fromList . M.keys . added) ] eqCoverageInv :: String -> [Partition -> Map] -> Invariants (Map, Partition) eqCoverageInv msg fs = sat $ do (s, p) <- cause doc $ msg ++ "\n\n" doc $ show p return $ (M.unions . map ($ p)) fs == s addedInv :: Invariants (Diff, Map) addedInv = sat $ do d <- fst <$> cause doc "added are the modulepath-set difference of incoming minus original" (==) (M.difference (incoming d) (original d)) . snd <$> cause -- TODO members of displacing either differ by property content, -- by file-path, or both displacingInv :: Invariants (Map, Map) displacingInv = sat $ do incoming_map <- fst <$> cause doc "displacing is a submap of incoming" (`M.isSubmapOf` incoming_map) . snd <$> cause equalInv :: Invariants (Diff, Map) equalInv = sat $ do (d, equal_map) <- cause doc "equal is a submap of original and incoming" doc $ "equal: " ++ show equal_map return $ (equal_map `M.isSubmapOf` incoming d) && (equal_map `M.isSubmapOf` original d) instance Arbitrary Diff where arbitrary = fromOriginal =<< fromMask M.empty =<< choose (0, 4) where -- | Produce a list of added sets, or sets that don't intersect on -- modulepath with any in the original set. fromMask :: Map -> Int -> Gen Map fromMask mask 0 = return mask fromMask mask n = do (p, satisfying_tm) <- maskSatisfying fromMask (M.insert (view modpath satisfying_tm) (p, view properties satisfying_tm) mask ) (n-1) where -- Note the suthThat predicate relies upon comparison of -- test module path, and this isn't very clear from -- appearances maskSatisfying :: Gen (FS.FilePath, TestModule) maskSatisfying = suchThat (toGenerated (choose ('a','z')) (M.keysSet mask)) -- =<< toModulePath (choose ('a', 'z'))) (\(_, tm) -> not . S.member (view modpath tm) $ M.keysSet mask) fromOriginal :: Map -> Gen Diff fromOriginal m = do intersecting <- choose (0, M.size m) equivalent <- choose (0, intersecting) changed <- choose (0, intersecting - equivalent) Diff m <$> toIncoming (M.toList m) equivalent changed where toIncoming :: [(ModulePath, (FS.FilePath, [P.Property]))] -> Int -> Int -> Gen Map toIncoming original_list equivalent changed = -- Partition the original into equivalent and changed and -- exclude the rest let equivalent_map = M.fromList . take equivalent $ original_list in do added_total <- choose (0,2) changed_map <- genChangedMap M.unions . (:[equivalent_map, changed_map]) <$> fromMask (equivalent_map `M.union` changed_map) added_total where genChangedMap = let subseq = take changed . drop equivalent $ original_list in do fs <- vectorOf (L.length subseq) mutation return . M.fromList $ zipWith (\f p -> over _2 f p) fs subseq -- Note, I could have used QuickCheck.Function here, but -- that is really overkill mutation :: Gen ((FS.FilePath, [P.Property]) -> (FS.FilePath, [P.Property])) mutation = oneof $ map return [over _1 pathMutation, over _2 propMutation] where pathMutation :: FS.FilePath -> FS.FilePath pathMutation p = FS.decodeString $ FS.encodeString "z/" ++ FS.encodeString p propMutation :: [P.Property] -> [P.Property] propMutation [] = [toProperty 0] propMutation (_:rest) = rest data Input_fromPartition = Input_fromPartition { masking :: Bool, original_map :: Map, partition :: Partition } deriving Show instance Arbitrary Input_fromPartition where arbitrary = do bool <- arbitrary :: Gen Bool diff <- suchThat (arbitrary :: Gen Diff) (not . M.null . original) return $ Input_fromPartition bool (original diff) (toPartition diff) -- TODO/FIXME cleanup/remove input type prop_fromPartition :: Input_fromPartition -> Property prop_fromPartition input = let result = runState (fromPartition (masking input) (partition input)) (original_map input) in runInvariants (result, input) inv_fromPartition where -- Note by precondition, -- all invariants on the Partition are held (disjoint, relations with -- imported etc). inv_fromPartition :: Invariants ((O.Ops, Map), Input_fromPartition) inv_fromPartition = do ops_invariants satcomp (\((_, ss), i) -> (ss,i)) suite_invariants where ops_invariants :: Invariants ((O.Ops, Map), Input_fromPartition) ops_invariants = let fromOpsLens lens t = set (_1 . _1) (view (_1 . _1 . lens) t) t addedRelations t = ((view (_1 . _1 . O.added) t, view (_1 . _2) t), partition . snd $ t ) removedRelations t = ((view (_1 . _1 . O.removed) t, view (_1 . _2) t), snd t) in do satcomp (toLabelled . fst . fst) disjointSetsInv satcomp addedRelations ops_added satcomp removedRelations ops_removed satcomp (fromOpsLens O.modified) ops_modified satcomp (fromOpsLens O.unmodified) ops_unmodified ops_added :: Invariants ((Map, Map), Partition) ops_added = do satcomp (\t -> (fst . fst $ t, added . snd $ t)) . sat $ do doc "added is equivalent to added in partition" uncurry (==) <$> cause satcomp fst . sat $ do doc "added is a submap of the suite mapping" uncurry M.isSubmapOf <$> cause toLabelled :: O.Ops -> [LabelledSet ModulePath] toLabelled ops = toSets ops [ ("added", S.fromList . M.keys . O._added) ,("removed", S.fromList . M.keys . O._removed) ,("modified", S.fromList . M.keys . O._modified) ,("unmodified", S.fromList . M.keys . O._unmodified) ] ops_removed :: Invariants ((Map, Map), Input_fromPartition) ops_removed = do sat $ do doc "removed is a subset of the original suite set" ((removed_map, _), i) <- cause return $ M.isSubmapOf removed_map (original_map i) satcomp fst . sat $ do doc "non-empty removed is not a subset of the final suite set" (removed_map, final_suite_map) <- cause return . (if M.null removed_map then id else not) $ M.isSubmapOf removed_map final_suite_map sat $ do ((removed_map, _), i) <- cause if masking i then do doc . build $ do "masking => removed is the difference: " " original - (equal of partition U displaced) " return $ (==) removed_map (M.difference (original_map i) (displaced `M.union` (equal . partition $ i)) ) else do doc "not masking => removed empty" return . M.null $ removed_map where displaced = M.intersection (original_map input) (displacing . partition $ input) ops_modified :: Invariants ((Map, Map), Input_fromPartition) ops_modified = do sat $ do doc "modified is a submap of original" ((modified_map, _), i) <- cause return $ M.isSubmapOf modified_map (original_map i) sat $ do doc . build $ do "modified is domain-equivalent to displacing of partition," "and for every equal relation (d1,d2) in a domain pairing," "s.t d1 -> v1 and d2 -> v2, v1 neq v2" ((modified_map, _), i) <- cause let displacing_map = displacing . partition $ i domain_eq = M.null $ M.difference modified_map displacing_map return . (&&) domain_eq . and . zipWith (/=) (M.elems modified_map) $ M.elems displacing_map ops_unmodified :: Invariants ((Map, Map), Input_fromPartition) ops_unmodified = do sat $ do doc "unmodified is a submap of original" ((unmodified_map, _), i) <- cause return $ M.isSubmapOf unmodified_map (original_map i) sat $ do ((unmodified_map, _), i) <- cause if masking i then do doc "masking => unmodified is only the equal set" return $ unmodified_map == (equal . partition $ i) else do doc "not masking => original - displaced-mapped of partition" return $ unmodified_map == M.difference (original_map i) ( M.intersection (original_map i) (displacing . partition $ i) ) suite_invariants :: Invariants (Map, Input_fromPartition) suite_invariants = sat $ do (final_suite, _) <- cause if masking input then do doc "the final suite is (equal U added U displacing)" return $ final_suite == minimum_expected else do doc . build $ do "the final suite is latent U (equal U added U displacing)" " where latent = original_set - displaced" " where displaced = domain mapped via displacing" return $ final_suite == M.union latent minimum_expected where latent = M.difference (original_map input) displaced displaced = M.intersection (original_map input) (displacing . partition $ input) minimum_expected = M.unions $ map ($ partition input) [equal, displacing, added] test_isomorphism :: Invariants (Tests, Tests) test_isomorphism = sat $ do doc "parsed tests are equal to those on filesystem" (written, parsed) <- cause if parsed /= written then do doc (msg parsed written) doc $ "\n written tests: \n" ++ (show . M.toList $ written) doc $ "\n parsed tests: \n" ++ (show . M.toList $ parsed) return False else return True where msg parsed written = if M.size parsed > M.size written then fromDifference "\n excessive parsed:\n " parsed written else fromDifference "\n unparsed:\n " written parsed fromDifference des l r = des ++ (show . M.toList $ M.difference l r) errors_isomorphism :: Invariants (Errors, Errors) errors_isomorphism = sat $ do doc "parsed errors are equal to those on filesystem" (written, parsed) <- cause if parsed /= written then do doc (msg parsed written) doc $ "\n written: \n" ++ (show . S.toList $ written) doc $ "\n parsed: \n" ++ (show . S.toList $ parsed) return False else return True where msg parsed written = if S.size parsed > S.size written then fromDifference "\n excessive parsed:\n " parsed written else fromDifference "\n unparsed:\n " written parsed fromDifference des l r = des ++ (show . S.toList $ S.difference l r) type Tests = M.Map ModulePath [String] type Errors = S.Set FS.FilePath -- FS is a Set of file-path(s), so it is sufficient to check for -- isomorphisms on the two path-disjoint sets forming its union, -- test files and non-sense files fromTestpathInv :: FS.FilePath -> Invariants (FS, (Map, [Error])) fromTestpathInv sandbox = do satcomp (uncurry to_test_isomorphism) test_isomorphism satcomp (uncurry to_errors_isomorphism) errors_isomorphism where to_test_isomorphism :: FS -> (Map, [Error]) -> (Tests, Tests) to_test_isomorphism fs (parsed_tests, _) = (,) (M.fromList . L.map (\tm -> (view modpath tm, map P.func $ view properties tm)) . M.elems . fst . M.mapEither id $ mapping fs) -- Written (M.map (map P.func . snd) parsed_tests) to_errors_isomorphism :: FS -> (Map, [Error]) -> (Errors, Errors) to_errors_isomorphism fs (_, parsed_errors) = (,) (S.map (FS.append sandbox) . M.keysSet . snd . M.mapEither id $ mapping fs) -- Given (S.fromList . map filepath $ parsed_errors) -- parsed -- FIXME/TODO still need to populate the directory with tests env :: String -> FS -> Env FS.FilePath env sandbox fs = Env { setup = do -- TODO add deleted state after completed createDirectory sandbox writeFS decoded_path fs return decoded_path , teardown = do removeDirectoryRecursive sandbox return () } where decoded_path = FS.decodeString sandbox filterIndicatives :: FS -> FS filterIndicatives = FS . M.filterWithKey (\k _ -> legalModSubpath . FS.encodeString $ k) . mapping prop_fromTestpath_isomorphism :: FS -> Property prop_fromTestpath_isomorphism fs = let sandbox = "./tmp/" in monadicIO $ do isomorphic <- run $ runEnv (env sandbox fs) toIsomorphic stop $ runInvariants (filterIndicatives fs, isomorphic) (fromTestpathInv $ FS.decodeString sandbox) where toIsomorphic :: FS.FilePath -> IO (Map, [Error]) toIsomorphic fp = do path <- toPath $ FS.encodeString fp runStateT (fromTestpath path) []
jfeltz/tasty-integrate
tests/Data/QcIntegrated.hs
bsd-2-clause
16,375
0
25
5,627
4,311
2,238
2,073
363
5
{-# LANGUAGE OverloadedStrings, GADTs #-} module View.Class where import Control.Applicative import Data.Monoid import Data.String import Model.Class import Model.Home import Model.Page import Web.Thermopolis.Clause import Types import View.Page import Web.Thermopolis.PageIdentity classPageClause :: (PageIdentity p f, p ~ Path, ContentReader f) => Page ClassContent -> f Clause classPageClause page = pageClause $ (\ ClassContent -> "... HA! HA HA...") <$> page
andygill/thermopolis
src/View/Class.hs
bsd-2-clause
572
0
8
166
121
69
52
14
1
{-# OPTIONS -fglasgow-exts #-} ----------------------------------------------------------------------------- {-| Module : QGL.hs Copyright : (c) David Harley 2010 Project : qtHaskell Version : 1.1.4 Modified : 2010-09-02 17:02:36 Warning : this file is machine generated - do not modify. --} ----------------------------------------------------------------------------- module Qtc.Enums.Opengl.QGL ( FormatOption, FormatOptions, eDoubleBuffer, fDoubleBuffer, eDepthBuffer, fDepthBuffer, eRgba, fRgba, eAlphaChannel, fAlphaChannel, eAccumBuffer, fAccumBuffer, eStencilBuffer, fStencilBuffer, eStereoBuffers, fStereoBuffers, eDirectRendering, fDirectRendering, eHasOverlay, fHasOverlay, eSampleBuffers, fSampleBuffers, eSingleBuffer, fSingleBuffer, eNoDepthBuffer, fNoDepthBuffer, eColorIndex, fColorIndex, eNoAlphaChannel, fNoAlphaChannel, eNoAccumBuffer, fNoAccumBuffer, eNoStencilBuffer, fNoStencilBuffer, eNoStereoBuffers, fNoStereoBuffers, eIndirectRendering, fIndirectRendering, eNoOverlay, fNoOverlay, eNoSampleBuffers, fNoSampleBuffers ) where import Qtc.Classes.Base import Qtc.ClassTypes.Core (QObject, TQObject, qObjectFromPtr) import Qtc.Core.Base (Qcs, connectSlot, qtc_connectSlot_int, wrapSlotHandler_int) import Qtc.Enums.Base import Qtc.Enums.Classes.Core data CFormatOption a = CFormatOption a type FormatOption = QEnum(CFormatOption Int) ieFormatOption :: Int -> FormatOption ieFormatOption x = QEnum (CFormatOption x) instance QEnumC (CFormatOption Int) where qEnum_toInt (QEnum (CFormatOption x)) = x qEnum_fromInt x = QEnum (CFormatOption x) withQEnumResult x = do ti <- x return $ qEnum_fromInt $ fromIntegral ti withQEnumListResult x = do til <- x return $ map qEnum_fromInt til instance Qcs (QObject c -> FormatOption -> IO ()) where connectSlot _qsig_obj _qsig_nam _qslt_obj _qslt_nam _handler = do funptr <- wrapSlotHandler_int slotHandlerWrapper_int stptr <- newStablePtr (Wrap _handler) withObjectPtr _qsig_obj $ \cobj_sig -> withCWString _qsig_nam $ \cstr_sig -> withObjectPtr _qslt_obj $ \cobj_slt -> withCWString _qslt_nam $ \cstr_slt -> qtc_connectSlot_int cobj_sig cstr_sig cobj_slt cstr_slt (toCFunPtr funptr) (castStablePtrToPtr stptr) return () where slotHandlerWrapper_int :: Ptr fun -> Ptr () -> Ptr (TQObject c) -> CInt -> IO () slotHandlerWrapper_int funptr stptr qobjptr cint = do qobj <- qObjectFromPtr qobjptr let hint = fromCInt cint if (objectIsNull qobj) then do when (stptr/=ptrNull) (freeStablePtr (castPtrToStablePtr stptr)) when (funptr/=ptrNull) (freeHaskellFunPtr (castPtrToFunPtr funptr)) else _handler qobj (qEnum_fromInt hint) return () data CFormatOptions a = CFormatOptions a type FormatOptions = QFlags(CFormatOptions Int) ifFormatOptions :: Int -> FormatOptions ifFormatOptions x = QFlags (CFormatOptions x) instance QFlagsC (CFormatOptions Int) where qFlags_toInt (QFlags (CFormatOptions x)) = x qFlags_fromInt x = QFlags (CFormatOptions x) withQFlagsResult x = do ti <- x return $ qFlags_fromInt $ fromIntegral ti withQFlagsListResult x = do til <- x return $ map qFlags_fromInt til instance Qcs (QObject c -> FormatOptions -> IO ()) where connectSlot _qsig_obj _qsig_nam _qslt_obj _qslt_nam _handler = do funptr <- wrapSlotHandler_int slotHandlerWrapper_int stptr <- newStablePtr (Wrap _handler) withObjectPtr _qsig_obj $ \cobj_sig -> withCWString _qsig_nam $ \cstr_sig -> withObjectPtr _qslt_obj $ \cobj_slt -> withCWString _qslt_nam $ \cstr_slt -> qtc_connectSlot_int cobj_sig cstr_sig cobj_slt cstr_slt (toCFunPtr funptr) (castStablePtrToPtr stptr) return () where slotHandlerWrapper_int :: Ptr fun -> Ptr () -> Ptr (TQObject c) -> CInt -> IO () slotHandlerWrapper_int funptr stptr qobjptr cint = do qobj <- qObjectFromPtr qobjptr let hint = fromCInt cint if (objectIsNull qobj) then do when (stptr/=ptrNull) (freeStablePtr (castPtrToStablePtr stptr)) when (funptr/=ptrNull) (freeHaskellFunPtr (castPtrToFunPtr funptr)) else _handler qobj (qFlags_fromInt hint) return () eDoubleBuffer :: FormatOption eDoubleBuffer = ieFormatOption $ 1 eDepthBuffer :: FormatOption eDepthBuffer = ieFormatOption $ 2 eRgba :: FormatOption eRgba = ieFormatOption $ 4 eAlphaChannel :: FormatOption eAlphaChannel = ieFormatOption $ 8 eAccumBuffer :: FormatOption eAccumBuffer = ieFormatOption $ 16 eStencilBuffer :: FormatOption eStencilBuffer = ieFormatOption $ 32 eStereoBuffers :: FormatOption eStereoBuffers = ieFormatOption $ 64 eDirectRendering :: FormatOption eDirectRendering = ieFormatOption $ 128 eHasOverlay :: FormatOption eHasOverlay = ieFormatOption $ 256 eSampleBuffers :: FormatOption eSampleBuffers = ieFormatOption $ 512 eSingleBuffer :: FormatOption eSingleBuffer = ieFormatOption $ 65536 eNoDepthBuffer :: FormatOption eNoDepthBuffer = ieFormatOption $ 131072 eColorIndex :: FormatOption eColorIndex = ieFormatOption $ 262144 eNoAlphaChannel :: FormatOption eNoAlphaChannel = ieFormatOption $ 524288 eNoAccumBuffer :: FormatOption eNoAccumBuffer = ieFormatOption $ 1048576 eNoStencilBuffer :: FormatOption eNoStencilBuffer = ieFormatOption $ 2097152 eNoStereoBuffers :: FormatOption eNoStereoBuffers = ieFormatOption $ 4194304 eIndirectRendering :: FormatOption eIndirectRendering = ieFormatOption $ 8388608 eNoOverlay :: FormatOption eNoOverlay = ieFormatOption $ 16777216 eNoSampleBuffers :: FormatOption eNoSampleBuffers = ieFormatOption $ 33554432 fDoubleBuffer :: FormatOptions fDoubleBuffer = ifFormatOptions $ 1 fDepthBuffer :: FormatOptions fDepthBuffer = ifFormatOptions $ 2 fRgba :: FormatOptions fRgba = ifFormatOptions $ 4 fAlphaChannel :: FormatOptions fAlphaChannel = ifFormatOptions $ 8 fAccumBuffer :: FormatOptions fAccumBuffer = ifFormatOptions $ 16 fStencilBuffer :: FormatOptions fStencilBuffer = ifFormatOptions $ 32 fStereoBuffers :: FormatOptions fStereoBuffers = ifFormatOptions $ 64 fDirectRendering :: FormatOptions fDirectRendering = ifFormatOptions $ 128 fHasOverlay :: FormatOptions fHasOverlay = ifFormatOptions $ 256 fSampleBuffers :: FormatOptions fSampleBuffers = ifFormatOptions $ 512 fSingleBuffer :: FormatOptions fSingleBuffer = ifFormatOptions $ 65536 fNoDepthBuffer :: FormatOptions fNoDepthBuffer = ifFormatOptions $ 131072 fColorIndex :: FormatOptions fColorIndex = ifFormatOptions $ 262144 fNoAlphaChannel :: FormatOptions fNoAlphaChannel = ifFormatOptions $ 524288 fNoAccumBuffer :: FormatOptions fNoAccumBuffer = ifFormatOptions $ 1048576 fNoStencilBuffer :: FormatOptions fNoStencilBuffer = ifFormatOptions $ 2097152 fNoStereoBuffers :: FormatOptions fNoStereoBuffers = ifFormatOptions $ 4194304 fIndirectRendering :: FormatOptions fIndirectRendering = ifFormatOptions $ 8388608 fNoOverlay :: FormatOptions fNoOverlay = ifFormatOptions $ 16777216 fNoSampleBuffers :: FormatOptions fNoSampleBuffers = ifFormatOptions $ 33554432
uduki/hsQt
Qtc/Enums/Opengl/QGL.hs
bsd-2-clause
7,292
0
18
1,333
1,720
910
810
202
1
{-#LANGUAGE DeriveGeneric #-} {-| Module : Name Description : Yaml parsing Copyright : License : GPL-2 Maintainer : theNerd247 Stability : experimental Portability : POSIX -} module Data.Serialization.Yaml ( readYamlFile ,writeYamlFile ) where import YabCommon import Data.Budget import Data.Serialization.Errors import qualified Data.Yaml as YAML import qualified Data.Aeson.Types as DAT data BadYAMLReadException = BadYAMLReadException FilePath deriving (Generic,Typeable,Read,Eq,Ord) instance Show BadYAMLReadException where show (BadYAMLReadException fpath) = show $ "Could not read YAML data from: " ++ fpath instance Exception BadYAMLReadException -- | Reads data from a yaml file. If no data could be read or an error occurs then the system exits. readYamlFile :: (YAML.FromJSON a, MonadIO m, MonadCatch m) => FilePath -> m a readYamlFile fpath = do r <- liftIO $ YAML.decodeFile fpath maybe (throwM $ BadYAMLReadException fpath) return $ r `catchAll` printEAndExit -- | write data to a yaml file. If an error occurs then the system exits writeYamlFile :: (YAML.ToJSON a, MonadIO m, MonadCatch m) => FilePath -> a -> m () writeYamlFile f = handleAll printEAndExit . liftIO . YAML.encodeFile f -- | TODO: add a way to support percentages of the income in the yaml files instance YAML.FromJSON Budget -- TODO: fix this quick hack as per issue #290 on github.com/bos/aeson instance YAML.ToJSON Budget instance YAML.ToJSON Account where toJSON = YAML.toJSON . accountAmount instance YAML.FromJSON Account where parseJSON v = do a <- YAML.parseJSON v return $ Account {accountAmount = a, accountEntries = []}
theNerd247/yab
src/Data/Serialization/Yaml.hs
bsd-3-clause
1,674
0
12
301
365
194
171
31
1
{-#LANGUAGE RankNTypes, CPP, Trustworthy #-} module Streaming ( -- * An iterable streaming monad transformer -- $stream Stream, -- * Constructing a 'Stream' on a given functor yields, effect, wrap, replicates, repeats, repeatsM, unfold, never, untilJust, streamBuild, delays, -- * Transforming streams maps, mapsM, mapped, distribute, groups, -- * Inspecting a stream inspect, -- * Splitting and joining 'Stream's splitsAt, takes, chunksOf, concats, intercalates, cutoff, -- period, -- periods, -- * Zipping, unzipping, separating and unseparating streams zipsWith, zips, unzips, interleaves, separate, unseparate, decompose, -- * Eliminating a 'Stream' mapsM_, run, streamFold, iterTM, iterT, destroy, -- * Base functor for streams of individual items Of (..), lazily, strictly, -- * ResourceT help bracketStream, -- * re-exports MFunctor(..), MMonad(..), MonadTrans(..), MonadIO(..), Compose(..), Sum(..), Identity(..), Alternative((<|>)), MonadThrow(..), MonadResource(..), MonadBase(..), ResourceT(..), runResourceT, #if MIN_VERSION_base(4,8,0) Bifunctor(..), #endif join, liftM, liftM2, liftA2, liftA3, void, (<>) ) where import Streaming.Internal import Streaming.Prelude import Control.Monad.Morph import Control.Monad import Data.Monoid ((<>)) import Control.Applicative import Control.Monad.Trans import Data.Functor.Compose import Data.Functor.Sum import Data.Functor.Identity import Data.Functor.Of import Control.Monad.Base import Control.Monad.Trans.Resource #if MIN_VERSION_base(4,8,0) import Data.Bifunctor #endif {- $stream The 'Stream' data type can be used to represent any effectful succession of steps arising in some monad. The form of the steps is specified by the first (\"functor\") parameter in @Stream f m r@. The monad of the underlying effects is expressed by the second parameter. This module exports combinators that pertain to that general case. Some of these are quite abstract and pervade any use of the library, e.g. > maps :: (forall x . f x -> g x) -> Stream f m r -> Stream g m r > mapped :: (forall x . f x -> m (g x)) -> Stream f m r -> Stream g m r > hoist :: (forall x . m x -> n x) -> Stream f m r -> Stream f n r -- from the MFunctor instance > concats :: Stream (Stream f m) m r -> Stream f m r (assuming here and thoughout that @m@ or @n@ satisfies a @Monad@ constraint, and @f@ or @g@ a @Functor@ constraint.) Others are surprisingly determinate in content: > chunksOf :: Int -> Stream f m r -> Stream (Stream f m) m r > splitsAt :: Int -> Stream f m r -> Stream f m (Stream f m r) > zipsWith :: (forall x y. f x -> g y -> h (x, y)) -> Stream f m r -> Stream g m r -> Stream h m r > intercalates :: Stream f m () -> Stream (Stream f m) m r -> Stream f m r > unzips :: Stream (Compose f g) m r -> Stream f (Stream g m) r > separate :: Stream (Sum f g) m r -> Stream f (Stream g) m r -- cp. partitionEithers > unseparate :: Stream f (Stream g) m r -> Stream (Sum f g) m r > groups :: Stream (Sum f g) m r -> Stream (Sum (Stream f m) (Stream g m)) m r One way to see that /any/ streaming library needs some such general type is that it is required to represent the segmentation of a stream, and to express the equivalents of @Prelude/Data.List@ combinators that involve 'lists of lists' and the like. See for example this <http://www.haskellforall.com/2013/09/perfect-streaming-using-pipes-bytestring.html post> on the correct expression of a streaming \'lines\' function. The module @Streaming.Prelude@ exports combinators relating to > Stream (Of a) m r where @Of a r = !a :> r@ is a left-strict pair. This expresses the concept of a 'Producer' or 'Source' or 'Generator' and easily inter-operates with types with such names in e.g. 'conduit', 'iostreams' and 'pipes'. -} {-| Map a stream to its church encoding; compare @Data.List.foldr@ This is the @safe_destroy@ exported by the @Internal@ module. Typical @FreeT@ operators can be defined in terms of @destroy@ e.g. > iterT :: (Functor f, Monad m) => (f (m a) -> m a) -> Stream f m a -> m a > iterT out stream = destroy stream out join return > iterTM :: (Functor f, Monad m, MonadTrans t, Monad (t m)) => (f (t m a) -> t m a) -> Stream f m a -> t m a > iterTM out stream = destroy stream out (join . lift) return > concats :: (Monad m, MonadTrans t, Monad (t m)) => Stream (t m) m a -> t m a > concats stream = destroy stream join (join . lift) return -}
michaelt/streaming
src/Streaming.hs
bsd-3-clause
4,789
0
5
1,186
358
250
108
80
0
import Control.Applicative boop = (*2) doop = (+10) -- function composition bip :: Integer -> Integer bip = boop . doop -- context is Functor bloop :: Integer -> Integer bloop = fmap boop doop -- context is Applicative bbop :: Integer -> Integer bbop = (+) <$> boop <*> doop duwop :: Integer -> Integer duwop = liftA2 (+) boop doop -- context is Monad boopDoop :: Integer -> Integer boopDoop = do a <- boop b <- doop return (a + b)
chengzh2008/hpffp
src/ch22-Reader/start.hs
bsd-3-clause
444
0
9
96
153
86
67
16
1
{-# LANGUAGE GeneralizedNewtypeDeriving, FlexibleContexts, MultiParamTypeClasses, FlexibleInstances #-} module Language.Mojito.Syntax.Types where import Prelude hiding ((<>)) import Data.List (intersperse) import Text.PrettyPrint.HughesPJClass -- Simple Types -- t = C t1 .. tn | a t1 .. tn data Simple = TyCon String | TyVar String | TyApp Simple Simple -- Or Simple [Simple], which would give an arity ? deriving (Show, Eq) instance Pretty Simple where pPrint (TyCon c) = text c pPrint (TyVar v) = text v pPrint (TyApp t1 t2) = text "(" <> pPrint t1 <+> pPrint t2 <> text ")" showSimple :: Simple -> String showSimple (TyCon a) = a showSimple (TyVar a) = a showSimple (TyApp t1 t2) = "(" ++ showSimple t1 ++ " " ++ showSimple t2 ++ ")" tc :: Simple -> [String] -- TODO nub the result, or use union instead of ++ -- tc is also Substitution.vars tc (TyCon s) = [s] tc (TyVar _) = [] tc (TyApp t1 t2) = tc t1 ++ tc t2 -- Occur check: tests if a type variable appears in a given -- type. occurs :: String -> Simple -> Bool occurs a (TyVar b) = a == b occurs _ (TyCon _) = False occurs a (TyApp t1 t2) = occurs a t1 || occurs a t2 -- Is it faster ? -- Occur check: tests if a type variable appears in a given -- type. -- occurs :: String -> Simple -> Bool -- occurs a t = a `elem` tv t tc' :: [Simple] -> [String] tc' = concatMap tc args :: Simple -> [Simple] args (TyCon _) = [] args (TyVar _) = [] args (TyApp t1 t2) = args t1 ++ [t2] nargs :: Simple -> Int nargs = length . args -- t == let (c,as) = cargs t -- in foldl TyApp c as cargs :: Simple -> (Simple,[Simple]) cargs t = cargs' t [] cargs' :: Simple -> [Simple] -> (Simple,[Simple]) cargs' t as = case t of TyCon _ -> (t,as) TyVar _ -> (t,as) TyApp t1 t2 -> cargs' t1 (t2 : as) -- fargs (-> a b c) = [a,b,c] fargs :: Simple -> [Simple] fargs (TyCon "->" `TyApp` a `TyApp` b) = a : fargs b fargs t = [t] -- Constrained types -- d = {oi : ti}.t data Constrained = Constrained [Constraint] Simple deriving (Show, Eq) showConstrained :: Constrained -> String showConstrained (Constrained [] t) = showSimple t showConstrained (Constrained cs t) = showConstraints cs ++ "." ++ showSimple t -- Constraint {o : t} where o is a term variable. data Constraint = Constraint String Simple deriving (Show, Eq) showConstraint :: Constraint -> String showConstraint (Constraint o t) = o ++ " : " ++ showSimple t showConstraints :: [Constraint] -> String showConstraints cs = "{" ++ concat (intersperse ", " $ map showConstraint cs) ++ "}" -- Quantified type. -- s = forall ai . d data Type = Type [String] Constrained deriving (Show, Eq) showType :: Type -> String showType (Type [] c) = showConstrained c showType (Type as c) = "forall " ++ concat (intersperse ", " as) ++ " . " ++ showConstrained c simple :: Type -> Simple simple (Type _ (Constrained _ t)) = t smpl :: Constrained -> Simple smpl (Constrained _ t) = t cstr :: Constrained -> [Constraint] cstr (Constrained c _) = c typed :: Simple -> Type typed t = Type [] $ Constrained [] t isSimple :: Type -> Bool isSimple (Type [] (Constrained [] _)) = True isSimple _ = False
noteed/mojito
Language/Mojito/Syntax/Types.hs
bsd-3-clause
3,139
0
10
656
1,139
600
539
71
3
{-# LANGUAGE OverloadedStrings #-} module RegexPattern where import Data.Text import Types -- ideally use regular expressions to create an item -- input: priority:description:MM/DD/YYYY regexPattern :: String regexPattern = "[0-9]+(:)[A-Za-z0-9 .!\"#$%()]+(:)[0-9]{1,2}/[0-9]{1,2}/[0-9]{2,4}" -- Used to read Items from file/stdin -- DOES NO PATTERN CHECKING - Left to regex. -- Assumes input is of proper format patternToItem :: String -> Item patternToItem itemString = Item (read p :: Int) desc date where list = Prelude.map unpack $ splitOn ":" (pack itemString) p = list !! 0 desc = list !! 1 date = list !! 2 -- Used to write Items to a file. itemToPattern :: Item -> String itemToPattern (Item p desc date) = (show p) ++ ":" ++ desc ++ ":" ++ date
frankhucek/Todo
app/RegexPattern.hs
bsd-3-clause
790
0
10
157
165
91
74
14
1
module Data.Params.ModInt where import Data.Params ------------------------------------------------------------------------------- -- general data family ModIntegral (modulus :: Param Nat) i type ModInt modulus = ModIntegral modulus Int type ModInteger modulus = ModIntegral modulus Integer ------------------------------------------------------------------------------- -- Static newtype instance ModIntegral (Static n) i = ModIntegral_Static i deriving (Read,Show,Eq,Ord) instance (KnownNat n, Integral i) => Num (ModIntegral (Static n) i) where (ModIntegral_Static a)+(ModIntegral_Static b) = ModIntegral_Static $ (a+b) `mod` n where n = fromIntegral $ natVal (Proxy::Proxy n) (ModIntegral_Static a)*(ModIntegral_Static b) = ModIntegral_Static $ (a*b) `mod` n where n = fromIntegral $ natVal (Proxy::Proxy n) abs = id signum = id fromInteger a = ModIntegral_Static $ (fromIntegral $ a `mod` n) where n = fromIntegral $ natVal (Proxy::Proxy n) negate (ModIntegral_Static a) = ModIntegral_Static $ (a-n) `mod` n where n = fromIntegral $ natVal (Proxy::Proxy n) ------------------------------------------------------------------------------- -- RunTime newtype instance ModIntegral RunTime i = ModIntegral_RunTime i deriving (Read,Show,Eq,Ord) mkParams ''ModIntegral instance ( Param_modulus (ModIntegral RunTime i) , Integral i ) => Num (ModIntegral RunTime i) where (ModIntegral_RunTime a)+(ModIntegral_RunTime b) = ModIntegral_RunTime $ (a+b) `mod` n where n = fromIntegral $ param_modulus (undefined:: (ModIntegral RunTime i)) (ModIntegral_RunTime a)*(ModIntegral_RunTime b) = ModIntegral_RunTime $ (a*b) `mod` n where n = fromIntegral $ param_modulus (undefined:: (ModIntegral RunTime i)) abs = id signum = id fromInteger a = ModIntegral_RunTime $ (fromIntegral $ a `mod` n) where n = fromIntegral $ param_modulus (undefined:: (ModIntegral RunTime i)) negate (ModIntegral_RunTime a) = ModIntegral_RunTime $ a-n `mod` n where n = fromIntegral $ param_modulus (undefined:: (ModIntegral RunTime i))
mikeizbicki/typeparams
src/Data/Params/ModInt.hs
bsd-3-clause
2,196
0
12
440
708
383
325
-1
-1
{-# LANGUAGE RecordWildCards #-} module Network.RMV.Post (sendRequest) where import Network.RMV.Types import Network.RMV.Output import Data.Char import Data.List import Data.Maybe import Network.Browser import Network.HTTP import Network.URI import System.Time rmvURI = URI { uriScheme = "http:" , uriAuthority = Just URIAuth { uriUserInfo = "" , uriRegName = "www.rmv.de" , uriPort = ":80" } , uriPath = "/auskunft/bin/jp/query.exe/dn" , uriQuery = "?L=vs_rmv&" , uriFragment = "" } reqBody :: Options -> IO String reqBody opts = do time <- getClockTime >>= toCalendarTime let curDate = intercalate "." . map (pad . show . ($time)) $ [ ctDay, (+1) . fromEnum . ctMonth, ctYear] curTime = ppTime (ctHour time * 60 + ctMin time) pad xs = replicate (2 - length xs) '0' ++ xs return $ intercalate "&" [ "queryPageDisplayed=yes" , "REQ0JourneyStopsN=-1" , "REQ0HafasNumCons1=1" , "REQ0HafasNumCons0=4" , "REQ0JourneyStopsS0A=255" , "REQ0JourneyStopsS0G=" ++ urlEncode (opStartPoint opts) , "REQ0JourneyStopsS0ID=" , "REQ0JourneyStopsZ0A=255" , "REQ0JourneyStopsZ0G=" ++ urlEncode (opEndPoint opts) , "REQ0JourneyStopsZ0ID=" , "existUnsharpSearch=yes" , "iER=no" , "REQ0HafasOptimize1=1" , "REQ0JourneyDate=" ++ urlEncode (fromMaybe curDate $ opDate opts) , "wDayExt0=Mo%7CDi%7CMi%7CDo%7CFr%7CSa%7CSo" , "REQ0JourneyTime=" ++ urlEncode (fromMaybe curTime $ opTime opts) , "REQ0HafasSearchForw=1" , "REQ0JourneyProductMask=1%3A11111111111111111" , "start=Verbindungen+suchen" ] userAgentMimicIE = "Mozilla/4.0 (compatible; MSIE 6.0; Windows NT 5.0; .NET CLR 1.1.4322; .NET CLR 2.0.50727; InfoPath.1)" expandAliases :: Options -> Options expandAliases opts@(Options{..}) = opts { opStartPoint = expand opStartPoint , opEndPoint = expand opEndPoint } where expand s = maybe s snd $ find ((~=s) . fst) aliases a ~= b = map toLower a == map toLower b -- | Sends a request with the specified options sendRequest :: Options -> IO String sendRequest opts = do req <- reqBody $ expandAliases opts rsp <- Network.Browser.browse $ do setAllowRedirects True setOutHandler $ const (return ()) request Request { rqURI = rmvURI , rqMethod = POST , rqHeaders = [ Header HdrContentLength $ show (length req) , Header HdrUserAgent userAgentMimicIE ] , rqBody = req } return . rspBody . snd $ rsp testOpts = defaultOptions { opStartPoint = "kopernikusplatz" , opEndPoint = "darmstadt, schloß" , opCount = Just 3 }
dschoepe/rmv-query
Network/RMV/Post.hs
bsd-3-clause
3,094
0
18
1,041
686
369
317
68
1
module Part1.Problem3 where -- -- The prime factors of 13195 are 5, 7, 13 and 29. -- -- What is the largest prime factor of the number 600851475143 ? -- https://stackoverflow.com/a/27464965/248948 squareRoot :: Integral t => t -> t squareRoot n | n > 0 = babylon n | n == 0 = 0 | n < 0 = error "Negative input" where babylon a | a > b = babylon b | otherwise = a where b = quot (a + quot n a) 2 primeSieve :: Int -> [Int] primeSieve n = filter (`notElem` sieve) [2..n] where sieve = [ i*i + j*i | i <- [2..squareRoot n] , j <- [0..((n - i*i) `div` i)] ] altPrimeSieve :: Integral t => t -> [t] altPrimeSieve n = filter isPrime [2..n] where isPrime p = all (\x -> p `mod` x /= 0) [2..(squareRoot p)] ceilSquareRoot :: Int -> Int ceilSquareRoot n = if s2 < n then s + 1 else s where s = squareRoot n s2 = s * s isSquare :: Int -> Bool isSquare n = s * s == n where s = squareRoot n fermatFactor :: Int -> Int fermatFactor n = factor' b2 a where a = ceilSquareRoot n b2 = a * a - n factor' b2 a = if not (isSquare b2) then let a' = a + 1 b2' = (a' * a') + n in factor' b2' a' else a - squareRoot b2 -- def trial_division(n): -- """Return a list of the prime factors for a natural number.""" -- if n < 2: -- return [] -- prime_factors = [] -- for p in prime_sieve(int(n**0.5)): -- if p*p > n: break -- while n % p == 0: -- prime_factors.append(p) -- n //= p -- if n > 1: -- prime_factors.append(n) -- return prime_factors trialDivision :: Integral t => t -> [t] trialDivision n | n < 2 = [] trialDivision n = trialDivision' n primes where primes = altPrimeSieve (squareRoot n) trialDivision' n [] | n > 1 = [n] | otherwise = [] trialDivision' n (p:ps) | p*p > n = trialDivision' n [] | otherwise = case factor n p of (x:xs) -> xs ++ trialDivision' x ps factor :: Integral t => t -> t -> [t] factor x p = if x `mod` p == 0 then factor (x `div` p) p ++ [p] else [x] largestPrimeFactor :: Int -> Int largestPrimeFactor p = maximum $ trialDivision p problem3 :: Int problem3 = largestPrimeFactor 600851475143
c0deaddict/project-euler
src/Part1/Problem3.hs
bsd-3-clause
2,277
0
15
700
830
433
397
56
3
{-#LANGUAGE ScopedTypeVariables, TypeOperators, FlexibleContexts #-} module Main where import Bindings.DC1394.Camera import Bindings.DC1394.Types import Canny import System.Environment import qualified Data.Array.Accelerate as A import Data.Array.Accelerate hiding (fromIntegral,fst,(++)) import qualified Data.Array.Accelerate.CUDA as CUDA import qualified Data.Array.Accelerate.IO as A import qualified Data.Array.Accelerate.IO.Firewire as A import Data.Array.Accelerate.Interpreter import qualified Data.Array.Repa as Repa import qualified Data.Array.Repa.IO.BMP as Repa import Data.Enumerator (run_) import Data.Maybe main :: IO () main = do numFrames <- (read . head) <$> getArgs withFirewireCamera ISO_400 Mode_640x480_RGB8 Rate_30 4 defaultFlags $ \camera -> do -- computation on one frame Right frame <- A.getFrame camera 640 480 let rgbaImg = (A.map A.packRGBA32 . A.flatten3dTo2d 480 640 . use) frame computation = (A.map A.rgba32OfLuminance .fst . canny 0.3 0.5) rgbaImg newImg = CUDA.run computation A.writeImageToBMP "out_accel.bmp" newImg -- computation on frame sequence let computation :: A.RGBImageDIM3 -> Acc (Array DIM2 A.RGBA32) computation = A.map A.rgba32OfLuminance -- convert graycale to RGB image . fst -- get the resulting grayscale image . canny 0.7 0.9 -- compute Canny edge detection . A.map A.packRGBA32 -- convert to 2D RGB (R,G,B) triple . A.flipV -- flip it in preparation for BMP file output . A.flatten3dTo2d 480 640 -- flatten 3D (Word8) image to 2D (Word8,Word8,Word8) image . use -- makes frame available for processing run_ $ A.withFrames camera 640 480 numFrames $ \frame frameCount -> do A.writeImageToBMP (show frameCount ++ "_out.bmp") (CUDA.run (computation frame)) -- run on the GPU
robstewart57/firewire-image-io
examples/accelerate.hs
bsd-3-clause
2,075
1
21
577
448
250
198
37
1
{-# OPTIONS_GHC -fno-warn-missing-signatures #-} {-# OPTIONS_GHC -fno-warn-type-defaults #-} module TestCodecovHaskellUtil where import Test.HUnit import Trace.Hpc.Codecov.Util testMapFirst = "mapFirst" ~: [ mapFirst (+ 1) [] @?= [], mapFirst (+ 1) [2] @?= [3], mapFirst (+ 1) [2, 3] @?= [3, 3], mapFirst (+ 1) [2, 3, 5] @?= [3, 3, 5]] testMapLast = "mapLast" ~: [ mapLast (+ 1) [] @?= [], mapLast (+ 1) [2] @?= [3], mapLast (+ 1) [2, 3] @?= [2, 4], mapLast (+ 1) [2, 3, 5] @?= [2, 3, 6]] testSubSeq = "subSeq" ~: [ subSeq 0 0 [] @?= ([] :: [Int]), subSeq 0 0 [2] @?= [], subSeq 0 1 [2] @?= [2], subSeq 0 2 [2] @?= [2], subSeq 1 1 [2] @?= [], subSeq 1 2 [2] @?= [], subSeq 0 0 [2, 3] @?= [], subSeq 0 1 [2, 3] @?= [2], subSeq 0 2 [2, 3] @?= [2, 3], subSeq 0 3 [2, 3] @?= [2, 3], subSeq 1 1 [2, 3] @?= [], subSeq 1 2 [2, 3] @?= [3], subSeq 1 3 [2, 3] @?= [3], subSeq 0 2 [2, 3] @?= [2, 3], subSeq 1 3 [2, 3, 5] @?= [3, 5], subSeq 0 3 [2, 3, 5] @?= [2, 3, 5]] testSubSubSeq = "subSubSeq" ~: [ subSubSeq 0 0 [[]] @?= ([[]] :: [[Int]]), subSubSeq 0 0 [[2]] @?= [[]], subSubSeq 0 1 [[2]] @?= [[2]], subSubSeq 0 0 [[2, 3]] @?= [[]], subSubSeq 0 1 [[2, 3]] @?= [[2]], subSubSeq 0 2 [[2, 3]] @?= [[2, 3]], subSubSeq 1 1 [[2, 3]] @?= [[]], subSubSeq 1 2 [[2, 3]] @?= [[3]], subSubSeq 1 3 [[2, 3]] @?= [[3]], subSubSeq 0 2 [[2, 3]] @?= [[2, 3]], subSubSeq 1 3 [[2, 3, 5]] @?= [[3, 5]], subSubSeq 0 3 [[2, 3, 5]] @?= [[2, 3, 5]], subSubSeq 0 0 [[2, 3], [5, 7]] @?= [[2, 3], []], subSubSeq 0 1 [[2, 3], [5, 7]] @?= [[2, 3], [5]], subSubSeq 0 2 [[2, 3], [5, 7]] @?= [[2, 3], [5, 7]], subSubSeq 1 0 [[2, 3], [5, 7]] @?= [[3], []], subSubSeq 1 1 [[2, 3], [5, 7]] @?= [[3], [5]], subSubSeq 1 2 [[2, 3], [5, 7]] @?= [[3], [5, 7]], subSubSeq 2 0 [[2, 3], [5, 7]] @?= [[], []], subSubSeq 2 1 [[2, 3], [5, 7]] @?= [[], [5]], subSubSeq 2 2 [[2, 3], [5, 7]] @?= [[], [5, 7]]] testGroupByIndex = "groupByIndex" ~: [ groupByIndex 0 [(0, 2)] @?= [], groupByIndex 1 [(0, 2)] @?= [[2]], groupByIndex 2 [(0, 2)] @?= [[2], []], groupByIndex 0 [(1, 2)] @?= [], groupByIndex 1 [(1, 2)] @?= [[]], groupByIndex 2 [(1, 2)] @?= [[], [2]], groupByIndex 3 [(1, 2)] @?= [[], [2], []], groupByIndex 0 [(0, 2), (0, 3)] @?= [], groupByIndex 1 [(0, 2), (0, 3)] @?= [[3, 2]], groupByIndex 1 [(0, 2), (1, 3)] @?= [[2]], groupByIndex 1 [(1, 2), (1, 3)] @?= [[]], groupByIndex 2 [(0, 2), (0, 3)] @?= [[3, 2], []], groupByIndex 2 [(0, 2), (1, 3)] @?= [[2], [3]], groupByIndex 2 [(1, 2), (1, 3)] @?= [[], [3, 2]], groupByIndex 3 [(0, 2), (0, 3)] @?= [[3, 2], [], []], groupByIndex 3 [(0, 2), (1, 3)] @?= [[2], [3], []], groupByIndex 3 [(1, 2), (1, 3)] @?= [[], [3, 2], []], groupByIndex 5 [(0, 2), (2, 5), (2, 3), (4, 13), (4, 11), (4, 7)] @?= [[2], [], [3, 5], [], [7, 11, 13]]] testUtil = "Util" ~: [ testMapFirst, testMapLast, testSubSeq, testSubSubSeq, testGroupByIndex]
guillaume-nargeot/codecov-haskell
test/TestCodecovHaskellUtil.hs
bsd-3-clause
3,108
0
10
856
2,166
1,290
876
79
1
module View.Talk where import Data.Aeson.Text (encodeToLazyText) import qualified Data.Text.Lazy as LT import Database.Persist import Lucid import Lucid.Base (makeAttribute) import Models.Talk (getTalkBySlug) import RIO import Types import View.Bundle import View.Error import View.Layout (layout) getTalkH :: Text -> AppM (Html ()) getTalkH slug = do mTalk <- lift $ getTalkBySlug slug case mTalk of Nothing -> get404H Just entity@(Entity _ talk) -> do bundle <- includeBundle TalkBundle layout ( do title_ $ toHtml $ talkName talk meta_ [ makeAttribute "property" "og:description" , name_ "description" , content_ $ talkDescription talk ] meta_ [ makeAttribute "property" "og:url" , content_ $ "https://ted2srt.org/talks/" <> talkSlug talk ] meta_ [ makeAttribute "property" "og:title" , content_ $ talkName talk ] meta_ [ makeAttribute "property" "og:type" , content_ "article" ] meta_ [ makeAttribute "property" "og:image" , content_ $ talkImage talk ] script_ $ LT.toStrict $ "window.TALK = " <> encodeToLazyText (entityIdToJSON entity) ) bundle
rnons/ted2srt
backend/src/View/Talk.hs
bsd-3-clause
1,496
0
19
590
342
172
170
41
2
{-# LANGUAGE NoImplicitPrelude #-} {-# LANGUAGE OverloadedStrings #-} module Test.Mismi.Amazonka ( sendMultipart , withMultipart , newMultipart ) where import Control.Monad.Catch import Control.Monad.Reader (ask) import Control.Monad.Trans.Resource import Data.Text as T import Data.Text.Encoding (encodeUtf8) import Mismi.S3 import qualified Mismi.S3.Amazonka as A import Mismi.S3.Internal import P import Test.Mismi.S3 import Test.QuickCheck import Test.QuickCheck.Instances () import X.Control.Monad.Trans.Either sendMultipart :: Text -> Address -> Int -> Text -> AWS () sendMultipart t a i ui = do let req = f' A.uploadPart a i ui (A.toBody $ encodeUtf8 t) void $ A.send req withMultipart :: Testable a => (Address -> Text -> AWS a) -> Property withMultipart f = testAWS $ do a <- newAddress awsBracket (createMultipartUpload a) (abortMultipart' a) (f a) newMultipart :: AWS (Address, Text) newMultipart = do a <- newAddress r <- createMultipartUpload a e <- ask void $ register (eitherT throwM pure . runAWS e $ abortMultipart' a r) void $ register (eitherT throwM pure . runAWS e $ listRecursively a >>= mapM_ delete >> delete a) pure $ (a, r)
ambiata/mismi
mismi-s3/test/Test/Mismi/Amazonka.hs
bsd-3-clause
1,315
0
14
334
412
218
194
36
1
{- TestVersion.hs (adapted from test_version.c in freealut) Copyright (c) Sven Panne 2005 <sven.panne@aedion.de> This file is part of the ALUT package & distributed under a BSD-style license See the file libraries/ALUT/LICENSE -} import Control.Monad ( unless ) import Sound.ALUT import System.Exit ( exitFailure ) import System.IO ( hPutStrLn, stderr ) -- This program checks that the version of OpenAL in the library agrees with the -- header file we're compiled against. main :: IO () main = withProgNameAndArgs runALUT $ \_progName _args -> do av <- get alutVersion unless (av == alutAPIVersion) $ do hPutStrLn stderr ("WARNING: The ALUT library is version " ++ av ++ ".x but the ALUT binding says it's " ++ alutAPIVersion ++ ".x!") exitFailure hPutStrLn stderr ("The ALUT library is at version " ++ av ++ ".x")
FranklinChen/hugs98-plus-Sep2006
packages/ALUT/examples/TestSuite/TestVersion.hs
bsd-3-clause
870
0
17
186
147
76
71
12
1
-- Copyright (c) 2016-present, Facebook, Inc. -- All rights reserved. -- -- This source code is licensed under the BSD-style license found in the -- LICENSE file in the root directory of this source tree. An additional grant -- of patent rights can be found in the PATENTS file in the same directory. {-# LANGUAGE GADTs #-} {-# LANGUAGE OverloadedStrings #-} module Duckling.Numeral.ZH.Rules ( rules ) where import qualified Data.HashMap.Strict as HashMap import Data.Maybe import qualified Data.Text as Text import Prelude import Data.String import Duckling.Dimensions.Types import Duckling.Numeral.Helpers import Duckling.Numeral.Types (NumeralData (..)) import qualified Duckling.Numeral.Types as TNumeral import Duckling.Regex.Types import Duckling.Types ruleInteger5 :: Rule ruleInteger5 = Rule { name = "integer (0..10)" , pattern = [ regex "(\x3007|\x96f6|\x4e00|\x4e8c|\x4e24|\x5169|\x4e09|\x56db|\x4e94|\x516d|\x4e03|\x516b|\x4e5d|\x5341)(\x4e2a|\x500b)?" ] , prod = \tokens -> case tokens of (Token RegexMatch (GroupMatch (match:_)):_) -> HashMap.lookup match integerMap >>= integer _ -> Nothing } integerMap :: HashMap.HashMap Text.Text Integer integerMap = HashMap.fromList [ ( "\x3007", 0 ) , ( "\x96f6", 0 ) , ( "\x4e00", 1 ) , ( "\x5169", 2 ) , ( "\x4e24", 2 ) , ( "\x4e8c", 2 ) , ( "\x4e09", 3 ) , ( "\x56db", 4 ) , ( "\x4e94", 5 ) , ( "\x516d", 6 ) , ( "\x4e03", 7 ) , ( "\x516b", 8 ) , ( "\x4e5d", 9 ) , ( "\x5341", 10 ) ] ruleNumeralsPrefixWithNegativeOrMinus :: Rule ruleNumeralsPrefixWithNegativeOrMinus = Rule { name = "numbers prefix with -, negative or minus" , pattern = [ regex "-|\x8d1f\\s?|\x8ca0\\s?" , dimension Numeral ] , prod = \tokens -> case tokens of (_:Token Numeral nd:_) -> double (TNumeral.value nd * (-1)) _ -> Nothing } ruleIntegerNumeric :: Rule ruleIntegerNumeric = Rule { name = "integer (numeric)" , pattern = [ regex "(\\d{1,18})" ] , prod = \tokens -> case tokens of (Token RegexMatch (GroupMatch (match:_)):_) -> do v <- parseInt match integer $ toInteger v _ -> Nothing } ruleDecimalWithThousandsSeparator :: Rule ruleDecimalWithThousandsSeparator = Rule { name = "decimal with thousands separator" , pattern = [ regex "(\\d+(,\\d\\d\\d)+\\.\\d+)" ] , prod = \tokens -> case tokens of (Token RegexMatch (GroupMatch (match:_)):_) -> parseDouble (Text.replace (Text.singleton ',') Text.empty match) >>= double _ -> Nothing } ruleDecimalNumeral :: Rule ruleDecimalNumeral = Rule { name = "decimal number" , pattern = [ regex "(\\d*\\.\\d+)" ] , prod = \tokens -> case tokens of (Token RegexMatch (GroupMatch (match:_)):_) -> parseDecimal True match _ -> Nothing } ruleNumeral :: Rule ruleNumeral = Rule { name = "<number>个" , pattern = [ dimension Numeral , regex "\x4e2a" ] , prod = \tokens -> case tokens of (token:_) -> Just token _ -> Nothing } ruleInteger3 :: Rule ruleInteger3 = Rule { name = "integer (20..90)" , pattern = [ numberBetween 2 10 , regex "\x5341" ] , prod = \tokens -> case tokens of (Token Numeral (NumeralData {TNumeral.value = v}):_) -> double $ v * 10 _ -> Nothing } ruleNumeralsSuffixesKMG :: Rule ruleNumeralsSuffixesKMG = Rule { name = "numbers suffixes (K, M, G)" , pattern = [ dimension Numeral , regex "([kmg])" ] , prod = \tokens -> case tokens of (Token Numeral (NumeralData {TNumeral.value = v}): Token RegexMatch (GroupMatch (match:_)): _) -> case Text.toLower match of "k" -> double $ v * 1e3 "m" -> double $ v * 1e6 "g" -> double $ v * 1e9 _ -> Nothing _ -> Nothing } ruleInteger4 :: Rule ruleInteger4 = Rule { name = "integer 21..99" , pattern = [ oneOf [70, 20, 60, 50, 40, 90, 30, 80] , numberBetween 1 10 ] , prod = \tokens -> case tokens of (Token Numeral (NumeralData {TNumeral.value = v1}): Token Numeral (NumeralData {TNumeral.value = v2}): _) -> double $ v1 + v2 _ -> Nothing } ruleInteger2 :: Rule ruleInteger2 = Rule { name = "integer (11..19)" , pattern = [ regex "\x5341" , numberBetween 1 10 ] , prod = \tokens -> case tokens of (_:Token Numeral (NumeralData {TNumeral.value = v}):_) -> double $ 10 + v _ -> Nothing } ruleIntegerWithThousandsSeparator :: Rule ruleIntegerWithThousandsSeparator = Rule { name = "integer with thousands separator ," , pattern = [ regex "(\\d{1,3}(,\\d\\d\\d){1,5})" ] , prod = \tokens -> case tokens of (Token RegexMatch (GroupMatch (match:_)): _) -> let fmt = Text.replace (Text.singleton ',') Text.empty match in parseDouble fmt >>= double _ -> Nothing } rules :: [Rule] rules = [ ruleDecimalNumeral , ruleDecimalWithThousandsSeparator , ruleInteger2 , ruleInteger3 , ruleInteger4 , ruleInteger5 , ruleIntegerNumeric , ruleIntegerWithThousandsSeparator , ruleNumeral , ruleNumeralsPrefixWithNegativeOrMinus , ruleNumeralsSuffixesKMG ]
rfranek/duckling
Duckling/Numeral/ZH/Rules.hs
bsd-3-clause
5,188
0
18
1,257
1,494
851
643
152
5
{-# LANGUAGE NamedFieldPuns #-} module RFB.CommandLoop (commandLoop) where import System.IO import qualified Data.ByteString.Lazy as BS import Data.Char import Data.Binary.Get import Data.Binary.Put import Data.Word import qualified RFB.ClientToServer as RFBClient import qualified RFB.Encoding as Encoding import qualified RFB.State as RFBState import Control.Monad.State type MyState a = StateT RFBState.RFBState IO a commandLoop :: Handle -> Int -> Int -> IO () commandLoop h width height= do runStateT (commandLoop1 h) (RFBState.initialState width height) return () commandLoop1 :: Handle -> MyState () commandLoop1 handle = do commandByte <- liftIO$ BS.hGet handle 1 let command = RFBClient.word8ToCommand $ runGet (do {x<-getWord8;return(x);}) commandByte byteStream <- liftIO $ BS.hGet handle (RFBClient.bytesToRead command) let commandData = RFBClient.getCommandData byteStream command --let commandData = RFBClient.parseCommandByteString byteStream command liftIO $ putStrLn $ "Client said -> " ++ (show command) liftIO $ putStrLn (show commandData) case command of RFBClient.SetPixelFormat -> processSetPixelFormat commandData RFBClient.SetEncodings -> liftIO $ processSetEncodings handle commandData RFBClient.FramebufferUpdate -> processFrameBufferUpdateRequest handle commandData RFBClient.PointerEvent -> processPointerEvent handle commandData RFBClient.ClientCutText -> processClientCutTextEvent handle commandData RFBClient.KeyEvent -> processKeyEvent handle commandData RFBClient.BadCommand -> liftIO $ putStrLn (show command) commandLoop1 handle processClientCutTextEvent handle (RFBClient.ClientCutTextData length) = do x <- liftIO $ BS.hGet handle length liftIO $ putStrLn (show x) processKeyEvent handle commandData = do return () processPointerEvent handle (RFBClient.PointerEventData buttonMask x y)= do state@(RFBState.RFBState d i p pend u) <- get let newState=if buttonMask == 1 then Encoding.putPixel state (x,y) else state put newState if pend && ((length u) > 0) then do liftIO $ putStrLn $ "Updating -> " ++ (show u) liftIO $ BS.hPutStr handle (updateRectangles state) liftIO $ hFlush handle put (RFBState.RFBState d i p False []) return () else return () processSetPixelFormat (RFBClient.SetPixelFormatData bitsPerPixel depth bigEndian trueColor redMax greenMax blueMax redShift greenShift blueShift) = do state <- get let newState=RFBState.setPixelFormat state bitsPerPixel bigEndian redMax greenMax blueMax redShift greenShift blueShift put newState return () processSetEncodings handle (RFBClient.SetEncodingsData count) = do x <- BS.hGet handle (4*count) let e = runGet (do {x<-getWord32le; return x}) x putStrLn (show e) return () fullScreen :: Int -> Int -> Int -> Int -> RFBState.RFBState -> BS.ByteString fullScreen x y width height state = runPut $ do putWord8 0 putWord8 0 putWord16be 1 putWord16be (fromIntegral x) putWord16be (fromIntegral y) putWord16be (fromIntegral width) putWord16be (fromIntegral height) putWord32be 0 Encoding.getImageByteString state x y width height return () updateRectangles state@(RFBState.RFBState _ _ (RFBState.PixelFormat bpp _ _ _ _ _ _ _) pend updateList) = runPut $ do putWord8 0 putWord8 0 putWord16be (fromIntegral (length updateList)) Encoding.encodeRectagles bpp updateList return () empty = runPut $ return () processFrameBufferUpdateRequest handle (RFBClient.FramebufferUpdateData incremental x y width height) = do -- this produces black state@(RFBState.RFBState d i p pend u) <- get let byteString=if incremental == 0 then fullScreen x y width height state else (if (length u > 0) then updateRectangles state else empty) --let newState = if incremental == 0 then state else RFBState.RFBState d i p pend [] let newState = if incremental == 0 then RFBState.RFBState d i p False [] else RFBState.RFBState d i p (if (length u > 0) then False else True) [] put newState (RFBState.RFBState _ _ _ newPend _) <- get liftIO $ putStrLn (show pend) liftIO $ putStrLn (show newPend) --liftIO $ putStrLn (show state) --liftIO $ putStrLn (show newState) liftIO $ BS.hPutStr handle byteString liftIO $ hFlush handle
ckkashyap/Chitra
RFB/CommandLoop.hs
bsd-3-clause
4,219
11
16
677
1,426
690
736
90
7
{-# LANGUAGE TypeFamilies, ViewPatterns #-} {-# OPTIONS_GHC -fno-warn-orphans #-} {-# LANGUAGE OverloadedStrings #-} ----------------------------------------------------------------------------- -- | -- Module : Program.View.WiredGeometryView -- Copyright : (c) Artem Chirkin -- License : MIT -- -- Maintainer : Artem Chirkin <chirkin@arch.ethz.ch> -- Stability : experimental -- -- -- ----------------------------------------------------------------------------- module Program.View.WiredGeometryView ( WiredGeometryView (..) )where import JsHs.TypedArray import JsHs.TypedArray.IO as JsHs import Data.Coerce (coerce) import JsHs.WebGL import SmallGL.Shader import Data.Geometry import Program.Model.WiredGeometry import Program.View data WiredGeometryView = WGView !WebGLBuffer !ShaderProgram !GLsizei instance Drawable WiredGeometry where type View WiredGeometry = WiredGeometryView createView gl (WiredGeometry _ _ size arr) = do buf <- createBuffer gl bindBuffer gl gl_ARRAY_BUFFER buf bufferData gl gl_ARRAY_BUFFER arr gl_STATIC_DRAW shProgram <- initShaders gl [(gl_FRAGMENT_SHADER, fragStaticMesh) ,(gl_VERTEX_SHADER, vertStaticMesh)] [] return $ WGView buf shProgram size drawInCurrContext vc@ViewContext{glctx = gl, curState = cs} (WiredGeometry _ (unpackV4 -> (r,g,b,a)) _ _) (WGView buf prog size) | size <= 1 = return () | otherwise = do enableVertexAttribArray gl ploc useProgram gl . programId $ prog bindBuffer gl gl_ARRAY_BUFFER buf uniformMatrix4fv gl (unifLoc prog "uProjM") False (projectArr vc) setIndex 0 (vView cs) (coerce $ modelViewArr vc) uniformMatrix4fv gl (unifLoc prog "uModelViewM") False (modelViewArr vc) uniform4f gl (unifLoc prog "uColor") (r*a) (g*a) (b*a) a vertexAttribPointer gl ploc 3 gl_FLOAT False 12 0 drawArrays gl gl_LINES 0 size disableVertexAttribArray gl ploc where ploc = attrLoc prog "aVertexPosition" updateView gl (WiredGeometry _ _ s arr) (WGView buf p _) = do bindBuffer gl gl_ARRAY_BUFFER buf bufferData gl gl_ARRAY_BUFFER arr gl_STATIC_DRAW return (WGView buf p s) deleteView gl _ (WGView buf _ _) = deleteBuffer gl buf fragStaticMesh :: String fragStaticMesh = unlines [ "precision mediump float;", "uniform vec4 uColor;", "varying vec3 vDist;", "void main(void) {", " gl_FragColor = clamp(3.0 - dot(vDist,vDist), 0.0, 1.0) * uColor;", "}"] vertStaticMesh :: String vertStaticMesh = unlines [ "precision mediump float;", "attribute vec3 aVertexPosition;", "uniform mat4 uModelViewM;", "uniform mat4 uProjM;", "varying vec3 vDist;", "void main(void) {", " vec4 globalPos = uModelViewM * vec4(aVertexPosition, 1.0);", " gl_Position = uProjM * globalPos;", " vDist = globalPos.xyz/(globalPos.w*200.0);", "}"]
achirkin/ghcjs-modeler
src/Program/View/WiredGeometryView.hs
bsd-3-clause
3,078
0
12
747
688
358
330
71
1
module PL0.SymbolTable ( SymEntry(..) , SymbolTable , emptySymbolTable ) where import PL0.Internal
LightAndLight/pl0-haskell
src/PL0/SymbolTable.hs
bsd-3-clause
116
0
5
29
26
17
9
5
0
{-# LANGUAGE TemplateHaskell #-} module Game.ClientRespawnT where import Control.Lens (makeLenses) import Linear (V3(..)) import Game.ClientPersistantT import Types makeLenses ''ClientRespawnT newClientRespawnT :: ClientRespawnT newClientRespawnT = ClientRespawnT { _crCoopRespawn = newClientPersistantT , _crEnterFrame = 0 , _crScore = 0 , _crCmdAngles = V3 0 0 0 , _crSpectator = False }
ksaveljev/hake-2
src/Game/ClientRespawnT.hs
bsd-3-clause
475
0
7
134
94
57
37
14
1
-- For everything {-# LANGUAGE GADTs #-} -- For type level flags, lists, strings {-# LANGUAGE DataKinds #-} {-# LANGUAGE KindSignatures #-} -- For Vinyl and universe polymorphism {-# LANGUAGE PolyKinds #-} -- For type-level lists {-# LANGUAGE TypeOperators #-} -- For unpacking existentials {-# LANGUAGE RankNTypes #-} -- For functions to apply to records {-# LANGUAGE TypeFamilies #-} -- For Aeson serialization {-# LANGUAGE OverloadedStrings #-} {-# LANGUAGE RankNTypes #-} {-# LANGUAGE FlexibleInstances #-} {-# LANGUAGE TemplateHaskell #-} {-# LANGUAGE ScopedTypeVariables #-} module Data.Master.Template ( -- * Data Kinds TemplateLevel(..) , Normalization(..) -- * Templating , Template(..) , TemplatesFor , TemplateFix(..) , unFixLevel , TemplateBox() , weakenTemplate , compileTemplateBox , compileTemplate , checkTemplate , checkTemplates -- * Lenses and prisms , templateFixNormalized , templateBox , templateFixBox , fixAnyLevelIso , Fixable (..) , _Or , _And , _Gt , _Lt , _Not , _Eq , _In ) where import Data.Aeson import Data.Aeson.Types import Control.Monad import Control.Applicative ((<$>), (<|>)) import qualified Control.Lens as L import Data.Vinyl.Core import Data.Vinyl.Functor import GHC.TypeLits -- | Take apart an or unOr :: Template Normalized Disjunction a -> [TemplateFix Normalized Conjunction a] unOr (Or fixes) = fixes -- | Take apart the existential 'FixAny' that permits violation of normalization constraints unFixAny :: (forall level' . Template Unnormalized level' a -> k) -> TemplateFix Unnormalized level a -> k unFixAny existentialKont (FixAny constraint) = existentialKont constraint -- | Take apart the 'FixLevel that prevents violation of normalization constraints unFixLevel :: TemplateFix Normalized level a -> Template Normalized level a unFixLevel (FixLevel c) = c -- | Functor transformer for template records type TemplatesFor norm level (f :: u -> *) = Compose (Template norm level) f -- | Check the templates for a record -- TODO: Have a real validation type checkTemplates :: Rec (TemplatesFor norm level f) fields -> Rec f fields -> Rec (Const Bool) fields checkTemplates constraintRec xRec = rmap (\(Compose constraint) -> Lift $ \x -> Const $ checkTemplate x constraint) constraintRec `rapply` xRec -- | Check whether a constraint is fulfilled checkTemplate :: a -> Template norm level a -> Bool checkTemplate _ Meh = True checkTemplate x (Not (FixAny template)) = not $ checkTemplate x template checkTemplate x (Not (FixLevel template)) = not $ checkTemplate x template checkTemplate x (Eq y) = x == y checkTemplate x (Lt y) = x < y checkTemplate x (Gt y) = x > y checkTemplate x (In y) = x `elem` y checkTemplate x (And []) = True checkTemplate x (And conjs@(FixAny _ : _)) = and $ map (unFixAny $ checkTemplate x) conjs checkTemplate x (And conjs@(FixLevel _ : _)) = and $ map (checkTemplate x . unFixLevel) conjs checkTemplate x (Or []) = False checkTemplate x (Or disjs@(FixAny _ : _)) = or $ map (unFixAny $ checkTemplate x) disjs checkTemplate x (Or disjs@(FixLevel _ : _)) = or $ map (checkTemplate x . unFixLevel) disjs -- | A constraint compiled by normalizing to horn clauses type CompiledTemplate = Template Normalized Disjunction -- | For enforcing normal form data TemplateLevel = Atom | Conjunction | Disjunction -- Flag for whether a type is normalized or not data Normalization = Unnormalized | Normalized -- When recurring, do we enforce normalization or not? data TemplateFix (norm :: Normalization) (level :: TemplateLevel) (a :: *) where FixAny :: Template Unnormalized level' a -> TemplateFix Unnormalized level a FixLevel :: Template Normalized level a -> TemplateFix Normalized level a -- | Isomorphism between a Normalized fix and the contained template templateFixNormalized :: L.Iso' (TemplateFix Normalized level a) (Template Normalized level a) templateFixNormalized = L.iso unFixLevel FixLevel instance (Eq a) => Eq (TemplateFix norm level a) where (==) = eqFixHelper eqFixHelper :: (Eq a) => TemplateFix norm level a -> TemplateFix norm level' a -> Bool eqFixHelper (FixAny a) (FixAny b) = eqHelper a b eqFixHelper (FixLevel a) (FixLevel b) = eqHelper a b eqFixHelper _ _ = False eqHelper :: (Eq a) => Template norm level' a -> Template norm level a -> Bool eqHelper Meh Meh = True eqHelper (Not a) (Not b) = a == b eqHelper (Eq a) (Eq b) = a == b eqHelper (Lt a) (Lt b) = a == b eqHelper (Gt a) (Gt b) = a == b eqHelper (In a) (In b) = a == b eqHelper (And a) (And b) = a == b eqHelper (Or a) (Or b) = a == b eqHelper _ _ = False -- | Templates data Template (norm :: Normalization) (level :: TemplateLevel) (a :: *) where Meh :: Template norm level a Not :: TemplateFix norm Atom a -> Template norm Atom a Eq :: (Eq a) => a -> Template norm Atom a Lt :: (Ord a) => a -> Template norm Atom a Gt :: (Ord a) => a -> Template norm Atom a In :: (Eq a) => [a] -> Template norm Atom a And :: [TemplateFix norm Atom a] -> Template norm Conjunction a Or :: [TemplateFix norm Conjunction a] -> Template norm Disjunction a instance (Eq a) => Eq (Template norm level a) where (==) = eqHelper -- | Convert a constraint from unnormalized form to normal form compileTemplate :: Template Unnormalized level a -> Template Normalized Disjunction a compileTemplate Meh = Meh compileTemplate (Not (FixAny Meh)) = Or [FixLevel $ And [FixLevel $ Not $ FixLevel $ Meh]] compileTemplate (Not (FixAny (Eq x))) = Or [FixLevel $ And [FixLevel $ Not $ FixLevel $ Eq x]] compileTemplate (Not (FixAny (Lt x))) = Or [FixLevel $ And [FixLevel $ Not $ FixLevel $ Lt x]] compileTemplate (Not (FixAny (Gt x))) = Or [FixLevel $ And [FixLevel $ Not $ FixLevel $ Gt x]] compileTemplate (Not (FixAny (In x))) = Or [FixLevel $ And [FixLevel $ Not $ FixLevel $ In x]] compileTemplate (Not (FixAny c)) = pushDownNots compileTemplate c compileTemplate (Eq x) = Or [FixLevel $ And [FixLevel $ Eq x]] compileTemplate (Lt x) = Or [FixLevel $ And [FixLevel $ Lt x]] compileTemplate (Gt x) = Or [FixLevel $ And [FixLevel $ Gt x]] compileTemplate (In x) = Or [FixLevel $ And [FixLevel $ In x]] compileTemplate (And fixes) = foldr distributeOrsAnd (Or [FixLevel $ And [FixLevel Meh]]) $ map (unFixAny compileTemplate) fixes compileTemplate (Or fixes) = Or $ concatMap (unFixAny $ unOr . compileTemplate) fixes -- | Push 'Nots' out to leaves pushDownNots :: (forall level' . Template Unnormalized level' a -> k) -> Template Unnormalized level a -> k pushDownNots k Meh = k $ Not $ FixAny Meh pushDownNots k (Not (FixAny c)) = k c pushDownNots k (Eq x) = k $ Not $ FixAny $ Eq x pushDownNots k (Lt x) = k $ Not $ FixAny $ Lt x pushDownNots k (Gt x) = k $ Not $ FixAny $ Gt x pushDownNots k (In x) = k $ Not $ FixAny $ In x pushDownNots k (And fixes) = k $ Or $ map (unFixAny $ pushDownNots FixAny) fixes pushDownNots k (Or fixes) = k $ And $ map (unFixAny $ pushDownNots FixAny) fixes -- | Distribute a conjunction of disjunctions of conjunctions to obtain a disjunction of conjunctions distributeOrsAnd :: Template Normalized Disjunction a -> Template Normalized Disjunction a -> Template Normalized Disjunction a distributeOrsAnd (Or andsLeft) (Or andsRight) = Or [FixLevel $ andAnds andLeft andRight | (FixLevel andLeft) <- andsLeft, (FixLevel andRight) <- andsRight] distributeOrsAnd Meh (Or andsRight) = Or andsRight distributeOrsAnd (Or andsLeft) Meh = Or andsLeft distributeOrsAnd Meh Meh = Meh andAnds :: Template Normalized Conjunction a -> Template Normalized Conjunction a -> Template Normalized Conjunction a andAnds (And leftFixes) (And rightFixes) = And $ leftFixes ++ rightFixes andAnds Meh (And rightFixes) = And rightFixes andAnds (And leftFixes) Meh = And leftFixes andAnds Meh Meh = Meh weakenTemplate :: Template norm level a -> TemplateBox a weakenTemplate Meh = TemplateBox Meh weakenTemplate (Eq x) = TemplateBox $ Eq x weakenTemplate (Not (FixAny t)) = case weakenTemplate t of TemplateBox t -> TemplateBox $ Not $ FixAny t weakenTemplate (Lt x) = TemplateBox $ Lt x weakenTemplate (Gt x) = TemplateBox $ Gt x weakenTemplate (In xs) = TemplateBox $ In xs weakenTemplate (And []) = TemplateBox $ And [] weakenTemplate (And ts@(FixAny _ : _)) = TemplateBox $ And $ ripTemplates (map (unFixAny weakenTemplate) ts) FixAny weakenTemplate (And ts@(FixLevel _ : _)) = TemplateBox $ And $ ripTemplates (map (weakenTemplate . unFixLevel) ts) FixAny weakenTemplate (Or []) = TemplateBox $ Or [] weakenTemplate (Or ts@(FixAny _ : _)) = TemplateBox $ Or $ ripTemplates (map (unFixAny weakenTemplate) ts) FixAny weakenTemplate (Or ts@(FixLevel _ : _)) = TemplateBox $ Or $ ripTemplates (map (weakenTemplate . unFixLevel) ts) FixAny compileTemplateBox :: TemplateBox a -> Template Normalized Disjunction a compileTemplateBox (TemplateBox t) = compileTemplate t instance (ToJSON a) => ToJSON (Template Normalized Disjunction a) where toJSON = toJSON . weakenTemplate instance (FromJSON a, Eq a, Ord a) => FromJSON (Template Normalized Disjunction a) where parseJSON = (compileTemplateBox <$>) . parseJSON data TemplateBox a = forall level. TemplateBox { _templateBox :: Template Unnormalized level a } -- | Isomorphism between a TemplateBox (used for encoding) and a TemplateFix Unnormalized (which can be deconstructed with prisms) templateBox :: L.Iso' (TemplateBox a) (TemplateFix Unnormalized level a) templateBox = L.iso (\(TemplateBox template) -> FixAny template) (\(FixAny template) -> TemplateBox template) data TemplateFixBox a = forall level. TemplateFixBox { _templateFixBox :: TemplateFix Unnormalized level a} templateFixBox :: L.Iso' (TemplateFixBox a) (TemplateFix Unnormalized level a) templateFixBox = L.iso (\(TemplateFixBox (FixAny template)) -> FixAny template) TemplateFixBox instance (ToJSON a) => ToJSON (TemplateFixBox a) where toJSON (TemplateFixBox (FixAny template)) = toJSON $ TemplateBox template instance (ToJSON a) => ToJSON (TemplateBox a) where toJSON (TemplateBox Meh) = object ["Meh" .= ()] toJSON (TemplateBox (Eq val)) = object ["Eq" .= val] toJSON (TemplateBox (Not t)) = object ["Not" .= TemplateFixBox t] toJSON (TemplateBox (Lt val)) = object ["Lt" .= val] toJSON (TemplateBox (Gt val)) = object ["Gt" .= val] toJSON (TemplateBox (In xs)) = object ["In" .= xs] toJSON (TemplateBox (And ts)) = object ["And" .= fmap TemplateFixBox ts] toJSON (TemplateBox (Or ts)) = object ["Or" .= fmap TemplateFixBox ts] instance (FromJSON a, Eq a, Ord a) => FromJSON (TemplateBox a) where parseJSON o = parseEq o <|> parseNot o <|> parseLt o <|> parseGt o <|> parseIn o <|> parseAnd o <|> parseOr o <|> parseMeh o parseEq, parseNot, parseLt, parseGt, parseIn, parseAnd, parseOr, parseMeh :: (FromJSON a, Ord a, Eq a) => Value -> Parser (TemplateBox a) parseEq (Object o) = do value <- (o .: "Eq") return $ TemplateBox $ Eq value parseEq _ = mzero parseNot (Object o) = do (TemplateBox template) <- o .: "Not" return $ TemplateBox $ Not $ FixAny $ template parseNot _ = mzero parseLt (Object o) = do value <- o .: "Lt" return $ TemplateBox $ Lt value parseLt _ = mzero parseGt (Object o) = do value <- o .: "Gt" return $ TemplateBox $ Gt value parseGt _ = mzero parseIn (Object o) = do xs <- o .: "In" return $ TemplateBox $ In xs parseIn _ = mzero parseAnd (Object o) = do xs <- o .: "And" return $ TemplateBox $ And $ ripTemplates xs FixAny parseAnd _ = mzero parseOr (Object o) = do xs <- o .: "Or" return $ TemplateBox $ Or $ ripTemplates xs FixAny parseOr _ = mzero parseMeh (Object o) = do () <- o .: "Meh" return $ TemplateBox $ Meh parseMeh _ = mzero ripTemplates :: [TemplateBox a] -> (forall level. Template Unnormalized level a -> k) -> [k] ripTemplates [] _ = [] ripTemplates ((TemplateBox t):ts) f = (f t):(ripTemplates ts f) -- | Necessary for the reconstruction side of prisms into TemplateBox. -- Simply propagate this constraint when using the prisms parametrically in the normalization variable. -- Instances are given for both cases. class Fixable (normalization :: Normalization) where fixFunction :: Template normalization level a -> TemplateFix normalization level a instance Fixable Normalized where fixFunction = FixLevel instance Fixable Unnormalized where fixFunction = FixAny -- | Isomorphism between TemplateFix unnormalized at differing levels. Use when applying a prism to Unnormalized boxes. fixAnyLevelIso :: L.Iso' (TemplateFix Unnormalized level1 a) (TemplateFix Unnormalized level2 a) fixAnyLevelIso = L.iso castFixAny castFixAny where castFixAny :: TemplateFix Unnormalized level1 a -> TemplateFix Unnormalized level2 a castFixAny (FixAny template) = FixAny template -- | Prism for getting/setting disjunctive constraints _Or :: (Fixable normalization) => L.Prism' (TemplateFix normalization Disjunction a) [TemplateFix normalization Conjunction a] _Or = L.prism' (fixFunction . Or) ripOr where ripOr :: TemplateFix normalization level a -> Maybe [TemplateFix normalization Conjunction a] ripOr (FixAny (Or xs)) = Just xs ripOr (FixLevel (Or xs)) = Just xs ripOr _ = Nothing -- | Prism for getting/setting conjunctive constraints _And :: (Fixable normalization) => L.Prism' (TemplateFix normalization Conjunction a) [TemplateFix normalization Atom a] _And = L.prism' (fixFunction . And) ripAnd where ripAnd :: TemplateFix normalization level a -> Maybe [TemplateFix normalization Atom a] ripAnd (FixAny (And xs)) = Just xs ripAnd (FixLevel (And xs)) = Just xs ripAnd _ = Nothing -- | Prism for getting/setting inverted constraints _Not :: (Fixable normalization) => L.Prism' (TemplateFix normalization Atom a) (TemplateFix normalization Atom a) _Not = L.prism' (fixFunction . Not) ripNot where ripNot :: (TemplateFix normalization level a) -> Maybe (TemplateFix normalization Atom a) ripNot (FixAny (Not t)) = Just t ripNot (FixLevel (Not t)) = Just t ripNot _ = Nothing -- | Prism for getting/setting equality constraints _Eq :: (Fixable normalization, Eq a) => L.Prism' (TemplateFix normalization Atom a) a _Eq = L.prism' (fixFunction . Eq) ripEq where ripEq :: (TemplateFix normalization level a) -> Maybe a ripEq (FixAny (Eq v)) = Just v ripEq (FixLevel (Eq v)) = Just v ripEq _ = Nothing -- | Prism for getting/setting less-than constraints _Lt :: (Fixable normalization, Ord a) => L.Prism' (TemplateFix normalization Atom a) a _Lt = L.prism' (fixFunction . Lt) ripLt where ripLt :: (TemplateFix normalization level a) -> Maybe a ripLt (FixAny (Lt v)) = Just v ripLt (FixLevel (Lt v)) = Just v ripLt _ = Nothing -- | Prism for getting/setting greater-than constraints _Gt :: (Fixable normalization, Ord a) => L.Prism' (TemplateFix normalization Atom a) a _Gt = L.prism' (fixFunction . Lt) ripGt where ripGt :: (TemplateFix normalization level a) -> Maybe a ripGt (FixAny (Gt v)) = Just v ripGt (FixLevel (Gt v)) = Just v ripGt _ = Nothing -- | Prism for getting/setting set membership constraints _In :: (Fixable normalization, Eq a) => L.Prism' (TemplateFix normalization Atom a) [a] _In = L.prism' (fixFunction . In) ripIn where ripIn :: (TemplateFix normalization level a) -> Maybe [a] ripIn (FixAny (In xs)) = Just xs ripIn (FixLevel (In xs)) = Just xs ripIn _ = Nothing
plow-technologies/template-service
master/src/Data/Master/Template.hs
bsd-3-clause
15,771
0
13
3,240
5,607
2,862
2,745
-1
-1